var/home/core/zuul-output/0000755000175000017500000000000015111240340014514 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111250364015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004432131515111250357017700 0ustar rootrootNov 25 05:40:57 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 05:40:57 crc restorecon[4565]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:57 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 05:40:58 crc restorecon[4565]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 05:40:58 crc kubenswrapper[4708]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.763126 4708 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765137 4708 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765156 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765160 4708 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765165 4708 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765169 4708 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765173 4708 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765177 4708 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765182 4708 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765186 4708 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765190 4708 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765193 4708 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765201 4708 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765206 4708 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765210 4708 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765214 4708 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765218 4708 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765222 4708 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765225 4708 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765229 4708 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765232 4708 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765237 4708 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765240 4708 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765243 4708 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765246 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765250 4708 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765253 4708 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765257 4708 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765260 4708 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765263 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765266 4708 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765278 4708 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765281 4708 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765285 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765288 4708 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765292 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765295 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765299 4708 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765304 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765308 4708 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765312 4708 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765315 4708 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765319 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765322 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765325 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765329 4708 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765332 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765337 4708 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765340 4708 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765343 4708 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765346 4708 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765350 4708 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765353 4708 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765356 4708 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765359 4708 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765363 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765366 4708 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765370 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765374 4708 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765380 4708 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765384 4708 feature_gate.go:330] unrecognized feature gate: Example Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765387 4708 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765391 4708 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765395 4708 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765398 4708 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765402 4708 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765405 4708 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765408 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765411 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765415 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765418 4708 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.765421 4708 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765489 4708 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765497 4708 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765505 4708 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765510 4708 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765531 4708 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765536 4708 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765541 4708 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765545 4708 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765549 4708 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765556 4708 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765561 4708 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765565 4708 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765568 4708 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765572 4708 flags.go:64] FLAG: --cgroup-root="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765576 4708 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765580 4708 flags.go:64] FLAG: --client-ca-file="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765584 4708 flags.go:64] FLAG: --cloud-config="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765588 4708 flags.go:64] FLAG: --cloud-provider="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765592 4708 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765598 4708 flags.go:64] FLAG: --cluster-domain="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765602 4708 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765606 4708 flags.go:64] FLAG: --config-dir="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765609 4708 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765613 4708 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765618 4708 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765622 4708 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765625 4708 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765629 4708 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765633 4708 flags.go:64] FLAG: --contention-profiling="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765637 4708 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765640 4708 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765644 4708 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765648 4708 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765652 4708 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765656 4708 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765660 4708 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765663 4708 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765666 4708 flags.go:64] FLAG: --enable-server="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765670 4708 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765674 4708 flags.go:64] FLAG: --event-burst="100" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765678 4708 flags.go:64] FLAG: --event-qps="50" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765682 4708 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765685 4708 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765689 4708 flags.go:64] FLAG: --eviction-hard="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765694 4708 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765698 4708 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765702 4708 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765706 4708 flags.go:64] FLAG: --eviction-soft="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765710 4708 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765714 4708 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765718 4708 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765722 4708 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765726 4708 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765729 4708 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765733 4708 flags.go:64] FLAG: --feature-gates="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765737 4708 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765740 4708 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765743 4708 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765747 4708 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765750 4708 flags.go:64] FLAG: --healthz-port="10248" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765754 4708 flags.go:64] FLAG: --help="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765757 4708 flags.go:64] FLAG: --hostname-override="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765760 4708 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765764 4708 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765768 4708 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765771 4708 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765774 4708 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765777 4708 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765781 4708 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765784 4708 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765787 4708 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765790 4708 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765794 4708 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765797 4708 flags.go:64] FLAG: --kube-reserved="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765800 4708 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765804 4708 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765807 4708 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765810 4708 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765813 4708 flags.go:64] FLAG: --lock-file="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765817 4708 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765820 4708 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765824 4708 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765831 4708 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765835 4708 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765838 4708 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765842 4708 flags.go:64] FLAG: --logging-format="text" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765845 4708 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765849 4708 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765852 4708 flags.go:64] FLAG: --manifest-url="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765856 4708 flags.go:64] FLAG: --manifest-url-header="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765860 4708 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765864 4708 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765867 4708 flags.go:64] FLAG: --max-pods="110" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765870 4708 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765874 4708 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765877 4708 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765881 4708 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765884 4708 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765888 4708 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765891 4708 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765899 4708 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765902 4708 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765906 4708 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765910 4708 flags.go:64] FLAG: --pod-cidr="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765913 4708 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765920 4708 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765923 4708 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765927 4708 flags.go:64] FLAG: --pods-per-core="0" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765931 4708 flags.go:64] FLAG: --port="10250" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765934 4708 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765938 4708 flags.go:64] FLAG: --provider-id="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765942 4708 flags.go:64] FLAG: --qos-reserved="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765945 4708 flags.go:64] FLAG: --read-only-port="10255" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765948 4708 flags.go:64] FLAG: --register-node="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765952 4708 flags.go:64] FLAG: --register-schedulable="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765956 4708 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765963 4708 flags.go:64] FLAG: --registry-burst="10" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765967 4708 flags.go:64] FLAG: --registry-qps="5" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765971 4708 flags.go:64] FLAG: --reserved-cpus="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765974 4708 flags.go:64] FLAG: --reserved-memory="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765978 4708 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765982 4708 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765985 4708 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765988 4708 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765991 4708 flags.go:64] FLAG: --runonce="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765995 4708 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.765998 4708 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766002 4708 flags.go:64] FLAG: --seccomp-default="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766005 4708 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766008 4708 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766011 4708 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766015 4708 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766018 4708 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766021 4708 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766025 4708 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766028 4708 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766031 4708 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766034 4708 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766038 4708 flags.go:64] FLAG: --system-cgroups="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766041 4708 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766046 4708 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766049 4708 flags.go:64] FLAG: --tls-cert-file="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766053 4708 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766057 4708 flags.go:64] FLAG: --tls-min-version="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766060 4708 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766064 4708 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766068 4708 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766072 4708 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766076 4708 flags.go:64] FLAG: --v="2" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766080 4708 flags.go:64] FLAG: --version="false" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766084 4708 flags.go:64] FLAG: --vmodule="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766089 4708 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766092 4708 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766175 4708 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766179 4708 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766183 4708 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766186 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766189 4708 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766192 4708 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766195 4708 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766198 4708 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766200 4708 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766203 4708 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766206 4708 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766209 4708 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766212 4708 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766215 4708 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766218 4708 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766221 4708 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766223 4708 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766226 4708 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766229 4708 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766232 4708 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766235 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766238 4708 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766241 4708 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766243 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766246 4708 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766250 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766253 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766256 4708 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766260 4708 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766263 4708 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766266 4708 feature_gate.go:330] unrecognized feature gate: Example Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766278 4708 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766282 4708 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766285 4708 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766288 4708 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766291 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766297 4708 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766300 4708 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766304 4708 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766307 4708 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766311 4708 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766315 4708 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766319 4708 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766322 4708 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766326 4708 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766329 4708 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766332 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766336 4708 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766339 4708 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766342 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766346 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766349 4708 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766352 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766355 4708 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766358 4708 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766361 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766364 4708 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766368 4708 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766371 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766374 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766377 4708 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766380 4708 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766383 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766387 4708 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766390 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766393 4708 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766396 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766399 4708 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766402 4708 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766404 4708 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.766407 4708 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.766417 4708 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.774606 4708 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.774635 4708 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774699 4708 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774710 4708 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774715 4708 feature_gate.go:330] unrecognized feature gate: Example Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774719 4708 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774723 4708 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774726 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774730 4708 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774733 4708 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774736 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774741 4708 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774744 4708 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774747 4708 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774750 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774754 4708 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774757 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774760 4708 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774763 4708 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774768 4708 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774774 4708 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774778 4708 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774782 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774785 4708 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774788 4708 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774793 4708 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774797 4708 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774801 4708 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774805 4708 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774810 4708 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774815 4708 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774820 4708 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774823 4708 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774826 4708 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774830 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774833 4708 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774836 4708 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774840 4708 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774843 4708 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774846 4708 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774849 4708 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774852 4708 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774856 4708 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774859 4708 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774862 4708 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774866 4708 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774869 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774874 4708 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774877 4708 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774881 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774884 4708 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774887 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774891 4708 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774894 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774897 4708 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774901 4708 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774904 4708 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774909 4708 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774913 4708 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774916 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774920 4708 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774923 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774928 4708 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774932 4708 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774936 4708 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774939 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774954 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774957 4708 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774961 4708 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774965 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774968 4708 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774971 4708 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.774975 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.774982 4708 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775093 4708 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775100 4708 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775104 4708 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775109 4708 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775112 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775116 4708 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775120 4708 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775124 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775127 4708 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775131 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775135 4708 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775139 4708 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775144 4708 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775149 4708 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775153 4708 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775157 4708 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775161 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775164 4708 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775168 4708 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775173 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775177 4708 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775181 4708 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775185 4708 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775189 4708 feature_gate.go:330] unrecognized feature gate: Example Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775193 4708 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775196 4708 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775200 4708 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775203 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775206 4708 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775209 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775213 4708 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775216 4708 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775219 4708 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775223 4708 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775227 4708 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775231 4708 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775235 4708 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775239 4708 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775243 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775246 4708 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775250 4708 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775253 4708 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775256 4708 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775260 4708 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775263 4708 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775266 4708 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775279 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775282 4708 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775285 4708 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775289 4708 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775292 4708 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775296 4708 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775299 4708 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775302 4708 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775305 4708 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775308 4708 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775312 4708 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775315 4708 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775319 4708 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775322 4708 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775325 4708 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775328 4708 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775331 4708 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775334 4708 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775337 4708 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775340 4708 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775343 4708 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775346 4708 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775350 4708 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775353 4708 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.775356 4708 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.775362 4708 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.776260 4708 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.778808 4708 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.778871 4708 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.779817 4708 server.go:997] "Starting client certificate rotation" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.779842 4708 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.779977 4708 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-29 01:13:24.784413147 +0000 UTC Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.780037 4708 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 91h32m26.004377893s for next certificate rotation Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.792148 4708 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.793531 4708 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.801636 4708 log.go:25] "Validated CRI v1 runtime API" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.818031 4708 log.go:25] "Validated CRI v1 image API" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.819237 4708 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.822154 4708 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-05-37-55-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.822173 4708 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:49 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm:{mountpoint:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm major:0 minor:42 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:50 fsType:tmpfs blockSize:0} overlay_0-43:{mountpoint:/var/lib/containers/storage/overlay/94b752e0a51c0134b00ddef6dc7a933a9d7c1d9bdc88a18dae4192a0d557d623/merged major:0 minor:43 fsType:overlay blockSize:0}] Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.838203 4708 manager.go:217] Machine: {Timestamp:2025-11-25 05:40:58.836740526 +0000 UTC m=+0.245573912 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2445406 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:62ba027d-f75d-44c0-93ed-f3312d588154 BootID:2c611337-7e15-4e6b-a515-a41d6885899c Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:49 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:50 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/var/lib/containers/storage/overlay-containers/75d81934760b26101869fbd8e4b5954c62b019c1cc3e5a0c9f82ed8de46b3b22/userdata/shm DeviceMajor:0 DeviceMinor:42 Capacity:65536000 Type:vfs Inodes:4108169 HasInodes:true} {Device:overlay_0-43 DeviceMajor:0 DeviceMinor:43 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:d1:13:df Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:enp3s0 MacAddress:fa:16:3e:d1:13:df Speed:-1 Mtu:1500} {Name:enp7s0 MacAddress:fa:16:3e:72:40:e4 Speed:-1 Mtu:1440} {Name:enp7s0.20 MacAddress:52:54:00:cf:0c:6b Speed:-1 Mtu:1436} {Name:enp7s0.21 MacAddress:52:54:00:ff:85:c2 Speed:-1 Mtu:1436} {Name:enp7s0.22 MacAddress:52:54:00:30:64:bf Speed:-1 Mtu:1436} {Name:eth10 MacAddress:5a:10:51:5b:36:d4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:2a:ae:db:17:64:98 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:65536 Type:Data Level:1} {Id:0 Size:65536 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:65536 Type:Data Level:1} {Id:1 Size:65536 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:65536 Type:Data Level:1} {Id:10 Size:65536 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:65536 Type:Data Level:1} {Id:11 Size:65536 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:65536 Type:Data Level:1} {Id:2 Size:65536 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:65536 Type:Data Level:1} {Id:3 Size:65536 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:65536 Type:Data Level:1} {Id:4 Size:65536 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:65536 Type:Data Level:1} {Id:5 Size:65536 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:65536 Type:Data Level:1} {Id:6 Size:65536 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:65536 Type:Data Level:1} {Id:7 Size:65536 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:65536 Type:Data Level:1} {Id:8 Size:65536 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:65536 Type:Data Level:1} {Id:9 Size:65536 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.838390 4708 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.838473 4708 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839013 4708 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839184 4708 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839208 4708 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839391 4708 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839400 4708 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839751 4708 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.839777 4708 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.840130 4708 state_mem.go:36] "Initialized new in-memory state store" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.840200 4708 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.842626 4708 kubelet.go:418] "Attempting to sync node with API server" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.842646 4708 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.842669 4708 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.842682 4708 kubelet.go:324] "Adding apiserver pod source" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.842694 4708 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.844882 4708 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.845570 4708 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.847206 4708 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.847640 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.847632 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.847734 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.847740 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848218 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848241 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848250 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848257 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848267 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848281 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848287 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848297 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848306 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848312 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848324 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848330 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.848773 4708 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.849085 4708 server.go:1280] "Started kubelet" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.849737 4708 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.849807 4708 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.849794 4708 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.850720 4708 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 05:40:58 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851584 4708 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851620 4708 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851909 4708 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851933 4708 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851804 4708 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-24 03:15:32.352059489 +0000 UTC Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.851996 4708 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 693h34m33.500065358s for next certificate rotation Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.852050 4708 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.852462 4708 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.852397 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="200ms" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.852664 4708 factory.go:55] Registering systemd factory Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.852685 4708 factory.go:221] Registration of the systemd container factory successfully Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.852643 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.852724 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.852396 4708 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 192.168.26.34:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b29745ffd4f75 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 05:40:58.849062773 +0000 UTC m=+0.257896159,LastTimestamp:2025-11-25 05:40:58.849062773 +0000 UTC m=+0.257896159,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.853106 4708 factory.go:153] Registering CRI-O factory Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.853121 4708 factory.go:221] Registration of the crio container factory successfully Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.854542 4708 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.854640 4708 factory.go:103] Registering Raw factory Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.856495 4708 manager.go:1196] Started watching for new ooms in manager Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.857669 4708 server.go:460] "Adding debug handlers to kubelet server" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.857768 4708 manager.go:319] Starting recovery of all containers Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865504 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865621 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865689 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865742 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865792 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865839 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865886 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865939 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.865991 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866040 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866090 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866151 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866201 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866259 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866328 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866379 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866431 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866481 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866617 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866708 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866771 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866826 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866875 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866930 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.866984 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867032 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867084 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867172 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867247 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867330 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867383 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867432 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867485 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867554 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867620 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867679 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867728 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867778 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867825 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867877 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867926 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.867974 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868023 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868070 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868123 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868173 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868220 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868279 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868334 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868382 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868437 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868486 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868559 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868613 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868664 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868723 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868775 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868825 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868875 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868924 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.868978 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869050 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869101 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869173 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869225 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869288 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869345 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869395 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869443 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869493 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869560 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869621 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869673 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869720 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869766 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869812 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869866 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869914 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.869963 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870016 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870087 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870161 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870237 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870305 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870362 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870411 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870464 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870537 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870591 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870639 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870687 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870743 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870829 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870886 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.870940 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.871786 4708 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.871887 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.871950 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872014 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872079 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872134 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872643 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872707 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872767 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872837 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872893 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.872945 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873019 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873085 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873151 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873213 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873264 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873344 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873402 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873452 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873502 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873573 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873623 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873672 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873788 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873860 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873917 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.873973 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874022 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874076 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874124 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874183 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874237 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874317 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874373 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874422 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874479 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874543 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874594 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874652 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874702 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874751 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874798 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874846 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874892 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874945 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.874994 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875043 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875089 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875140 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875196 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875245 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875321 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875372 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875434 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875385 4708 manager.go:324] Recovery completed Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875494 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875664 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875713 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875762 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875820 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875875 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875922 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.875999 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876048 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876094 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876150 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876198 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876245 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876306 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876356 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876402 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876456 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876505 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876578 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876628 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876686 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876773 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876830 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876878 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876927 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.876974 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877031 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877617 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877682 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877734 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877785 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877847 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877907 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.877959 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878015 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878068 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878117 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878173 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878223 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878285 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878337 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878386 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878441 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878493 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878561 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878614 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878662 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878719 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878771 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878820 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878867 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878915 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.878964 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879028 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879079 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879126 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879176 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879227 4708 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879296 4708 reconstruct.go:97] "Volume reconstruction finished" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.879344 4708 reconciler.go:26] "Reconciler: start to sync state" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.883467 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.888990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.889016 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.889025 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.889474 4708 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.889486 4708 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.889502 4708 state_mem.go:36] "Initialized new in-memory state store" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.890127 4708 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.892000 4708 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.892034 4708 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.892055 4708 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.892109 4708 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 05:40:58 crc kubenswrapper[4708]: W1125 05:40:58.892459 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.892502 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.893815 4708 policy_none.go:49] "None policy: Start" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.894217 4708 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.894243 4708 state_mem.go:35] "Initializing new in-memory state store" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937104 4708 manager.go:334] "Starting Device Plugin manager" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937256 4708 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937281 4708 server.go:79] "Starting device plugin registration server" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937618 4708 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937630 4708 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937754 4708 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937872 4708 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.937879 4708 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 05:40:58 crc kubenswrapper[4708]: E1125 05:40:58.943220 4708 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.992505 4708 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.992618 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.993416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.993454 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.993563 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.993787 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.993978 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.994096 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.994686 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.994706 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.994714 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995096 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995123 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995226 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995331 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995381 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995750 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995775 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995786 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995873 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.995971 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996006 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996550 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996560 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996563 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996594 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996862 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996942 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.996970 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997003 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997005 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997018 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997804 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997846 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997855 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997890 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997904 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.997913 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.998035 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.998082 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.998601 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.998622 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:58 crc kubenswrapper[4708]: I1125 05:40:58.998630 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.037793 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.038728 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.038763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.038773 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.038804 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.039233 4708 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.34:6443: connect: connection refused" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.052906 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="400ms" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.080974 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081113 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081324 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081429 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081545 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081621 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081689 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081758 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081824 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081911 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.081968 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.082042 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.082422 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.082670 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.082759 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183327 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183383 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183414 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183440 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183469 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183467 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183552 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183533 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183580 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183581 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183630 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183738 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183791 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183824 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183758 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183861 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183877 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183892 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183894 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183928 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183936 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183963 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183980 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.183998 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.184006 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.184038 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.184047 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.184106 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.239569 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.240457 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.240570 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.240636 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.240715 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.241131 4708 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.34:6443: connect: connection refused" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.313239 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: W1125 05:40:59.334488 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-251ab8ad2de3b8eefafa4d3f5d2541e658eccedb3ffeb752d5ae6eb58bdf1bf7 WatchSource:0}: Error finding container 251ab8ad2de3b8eefafa4d3f5d2541e658eccedb3ffeb752d5ae6eb58bdf1bf7: Status 404 returned error can't find the container with id 251ab8ad2de3b8eefafa4d3f5d2541e658eccedb3ffeb752d5ae6eb58bdf1bf7 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.337758 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.344968 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: W1125 05:40:59.356684 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-470bdafa6366ef45adb82b2a6de3cec5ebd154de16f53d57b277fea12224d694 WatchSource:0}: Error finding container 470bdafa6366ef45adb82b2a6de3cec5ebd154de16f53d57b277fea12224d694: Status 404 returned error can't find the container with id 470bdafa6366ef45adb82b2a6de3cec5ebd154de16f53d57b277fea12224d694 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.368837 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.371988 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 05:40:59 crc kubenswrapper[4708]: W1125 05:40:59.376892 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-e16961a95928f41b05e292ae36691830f5b87797f29ccc3b14012fa5549a8dad WatchSource:0}: Error finding container e16961a95928f41b05e292ae36691830f5b87797f29ccc3b14012fa5549a8dad: Status 404 returned error can't find the container with id e16961a95928f41b05e292ae36691830f5b87797f29ccc3b14012fa5549a8dad Nov 25 05:40:59 crc kubenswrapper[4708]: W1125 05:40:59.382065 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-e59d263c5c699c3ecf70744c5884940009fd55402014047715c01b90da75e74f WatchSource:0}: Error finding container e59d263c5c699c3ecf70744c5884940009fd55402014047715c01b90da75e74f: Status 404 returned error can't find the container with id e59d263c5c699c3ecf70744c5884940009fd55402014047715c01b90da75e74f Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.453847 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="800ms" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.642027 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.643096 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.643346 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.643358 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.643383 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.643798 4708 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.34:6443: connect: connection refused" node="crc" Nov 25 05:40:59 crc kubenswrapper[4708]: W1125 05:40:59.740647 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:59 crc kubenswrapper[4708]: E1125 05:40:59.740718 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.850328 4708 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.896612 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897" exitCode=0 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.896707 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.896814 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"470bdafa6366ef45adb82b2a6de3cec5ebd154de16f53d57b277fea12224d694"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.896946 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.897842 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.897884 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.897896 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.898553 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.898598 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1cb0e7deb768f2837139b50f0b60b1d60dcb4af41e34f6063bb29f0a501d7892"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.899344 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.900103 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"d4925eaa62dac51117085bae6732f6b65b1ba83ff0615e7869ce5c455081ee0d"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.900321 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.900069 4708 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="d4925eaa62dac51117085bae6732f6b65b1ba83ff0615e7869ce5c455081ee0d" exitCode=0 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.900950 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"251ab8ad2de3b8eefafa4d3f5d2541e658eccedb3ffeb752d5ae6eb58bdf1bf7"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902357 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902357 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902388 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902373 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.902400 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.904172 4708 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="874a6d7983c65f13276ab2f86fac9ecdfd4753473ca55f2c6b4cc4887bf17ba1" exitCode=0 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.904239 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"874a6d7983c65f13276ab2f86fac9ecdfd4753473ca55f2c6b4cc4887bf17ba1"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.904263 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"e59d263c5c699c3ecf70744c5884940009fd55402014047715c01b90da75e74f"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.904435 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905493 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905510 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905913 4708 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec" exitCode=0 Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905937 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.905955 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"e16961a95928f41b05e292ae36691830f5b87797f29ccc3b14012fa5549a8dad"} Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.906045 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.906568 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.906594 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:40:59 crc kubenswrapper[4708]: I1125 05:40:59.906603 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: W1125 05:41:00.204738 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:41:00 crc kubenswrapper[4708]: E1125 05:41:00.205231 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:41:00 crc kubenswrapper[4708]: E1125 05:41:00.254224 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="1.6s" Nov 25 05:41:00 crc kubenswrapper[4708]: W1125 05:41:00.280034 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:41:00 crc kubenswrapper[4708]: E1125 05:41:00.280093 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:41:00 crc kubenswrapper[4708]: W1125 05:41:00.354129 4708 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 192.168.26.34:6443: connect: connection refused Nov 25 05:41:00 crc kubenswrapper[4708]: E1125 05:41:00.354196 4708 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 192.168.26.34:6443: connect: connection refused" logger="UnhandledError" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.444269 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.445982 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.446012 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.446021 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.446044 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:41:00 crc kubenswrapper[4708]: E1125 05:41:00.446405 4708 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 192.168.26.34:6443: connect: connection refused" node="crc" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.910496 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.910550 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.910554 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.910566 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.911404 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.911432 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.911443 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913571 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913605 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913617 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913625 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913641 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.913693 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.914249 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.914275 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.914284 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.915820 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.915845 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.915856 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.915848 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.916598 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.916616 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.916624 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917408 4708 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="728c52f783d4e672f2704e9f8ff58f6680fd3dc668ecb2665584deae15fcd95a" exitCode=0 Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917445 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"728c52f783d4e672f2704e9f8ff58f6680fd3dc668ecb2665584deae15fcd95a"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917509 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917969 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.917993 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.919255 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9c91c363f72c1eec026fb1f559609070ac7660e311f094db141b407d7fbff2cf"} Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.919306 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.923878 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.923894 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:00 crc kubenswrapper[4708]: I1125 05:41:00.923902 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923580 4708 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7fd9815d2cdd4466082fa584507e7062673143f9316fda2af0c01bf85cdec4e0" exitCode=0 Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923667 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7fd9815d2cdd4466082fa584507e7062673143f9316fda2af0c01bf85cdec4e0"} Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923680 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923731 4708 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923744 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923765 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923804 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.923752 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.924906 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.924931 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.924940 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.924988 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.925007 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.925017 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.925280 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.925313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.925323 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.926265 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.926298 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:01 crc kubenswrapper[4708]: I1125 05:41:01.926309 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.046821 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.047899 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.047938 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.047951 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.047983 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.819899 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.931915 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"11845ff24344a660f66680ca8afa3479455f903c95d37580635ff77d557da6fc"} Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.931961 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"724654da2eeb1aa76f70b1e8e399b257734106e269847309d593bdfaa00bda45"} Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.931980 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"348ece851be55ceeccdcd9bfee4791e0528e3d2fd845222536e5ea65868086ac"} Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.931996 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a7a02af98daf012652d03bf368b0d53609ab2c2e16def5eb0b7f68c25bddd5af"} Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.932011 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.932049 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.932012 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2d1fa58225fd09603a4897a0654e2a69e7ae48e8c8ef06ec93b32a85ffcff688"} Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.931994 4708 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.932160 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933066 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933102 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933111 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933072 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933176 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933432 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:02 crc kubenswrapper[4708]: I1125 05:41:02.933445 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.129586 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.395350 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.395568 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.396465 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.396501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.396510 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.932754 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.934147 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.934274 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.934918 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.934943 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.934951 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.935105 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.935148 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:03 crc kubenswrapper[4708]: I1125 05:41:03.935162 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.618955 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.936182 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.936241 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937098 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937126 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937098 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937136 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937147 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:04 crc kubenswrapper[4708]: I1125 05:41:04.937280 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.021448 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.021585 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.022423 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.022455 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.022465 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.396360 4708 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.396416 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.714499 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.940435 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.941122 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.941152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:06 crc kubenswrapper[4708]: I1125 05:41:06.941161 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:07 crc kubenswrapper[4708]: I1125 05:41:07.537124 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 05:41:07 crc kubenswrapper[4708]: I1125 05:41:07.537290 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:07 crc kubenswrapper[4708]: I1125 05:41:07.538350 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:07 crc kubenswrapper[4708]: I1125 05:41:07.538462 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:07 crc kubenswrapper[4708]: I1125 05:41:07.538576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.281245 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.281356 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.282165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.282340 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.282502 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.285388 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:08 crc kubenswrapper[4708]: E1125 05:41:08.943295 4708 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.943344 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.946015 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.946040 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:08 crc kubenswrapper[4708]: I1125 05:41:08.946050 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:10 crc kubenswrapper[4708]: I1125 05:41:10.852225 4708 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 05:41:11 crc kubenswrapper[4708]: I1125 05:41:11.245672 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 05:41:11 crc kubenswrapper[4708]: I1125 05:41:11.245726 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 05:41:11 crc kubenswrapper[4708]: I1125 05:41:11.254510 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 05:41:11 crc kubenswrapper[4708]: I1125 05:41:11.254571 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.825614 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.825741 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.826572 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.826613 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.826622 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.828865 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.951678 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.984679 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.984727 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:12 crc kubenswrapper[4708]: I1125 05:41:12.984737 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.635879 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.636423 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.637212 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.637241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.637268 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.643789 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.954889 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.955619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.955655 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:14 crc kubenswrapper[4708]: I1125 05:41:14.955664 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.245322 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.246917 4708 trace.go:236] Trace[1595474983]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 05:41:03.122) (total time: 13124ms): Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[1595474983]: ---"Objects listed" error: 13124ms (05:41:16.246) Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[1595474983]: [13.124122273s] [13.124122273s] END Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.246944 4708 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.247985 4708 trace.go:236] Trace[477171174]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 05:41:02.640) (total time: 13607ms): Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[477171174]: ---"Objects listed" error: 13607ms (05:41:16.247) Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[477171174]: [13.607656526s] [13.607656526s] END Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.248002 4708 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.248113 4708 trace.go:236] Trace[1011917416]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 05:41:02.274) (total time: 13973ms): Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[1011917416]: ---"Objects listed" error: 13973ms (05:41:16.248) Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[1011917416]: [13.97349952s] [13.97349952s] END Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.248131 4708 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.249823 4708 trace.go:236] Trace[384554720]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 05:41:02.779) (total time: 13470ms): Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[384554720]: ---"Objects listed" error: 13470ms (05:41:16.249) Nov 25 05:41:16 crc kubenswrapper[4708]: Trace[384554720]: [13.470163421s] [13.470163421s] END Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.249848 4708 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.253316 4708 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.255143 4708 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280295 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280320 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280340 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280366 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280832 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57192->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.280855 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:57192->192.168.126.11:17697: read: connection reset by peer" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.281105 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.281144 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.288139 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.291599 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.293088 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.852728 4708 apiserver.go:52] "Watching apiserver" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.854747 4708 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855038 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-clxd5","openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855366 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855403 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855461 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.855545 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855601 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.855656 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855831 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855891 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.855956 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.856114 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.856714 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.856719 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.856735 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.857105 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.857680 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.857700 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.857780 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.857876 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.859618 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.859752 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.859781 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.862730 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.871008 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.877232 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.885570 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.891981 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.901467 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-lh62q"] Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.901755 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.902148 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.903945 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.904173 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.904208 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.904401 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.904513 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.914745 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.921867 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.928217 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.933770 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.939645 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.945403 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.951082 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.952544 4708 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956443 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956488 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956502 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956533 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956548 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956562 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956577 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956594 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956608 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956625 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956638 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956652 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956668 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956682 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956697 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956710 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956724 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956738 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956771 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956806 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956806 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956820 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956878 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956897 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956913 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956931 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956927 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956947 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956964 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956980 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956994 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.956994 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957009 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957025 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957038 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957052 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957066 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957079 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957092 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957095 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957110 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957125 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957139 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957154 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957167 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957181 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957187 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957196 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957210 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957221 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957223 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957262 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957289 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957306 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957320 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957333 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957361 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957378 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957390 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957404 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957419 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957441 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957455 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957485 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957498 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957513 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957544 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957559 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957574 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957605 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957618 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957635 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957655 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957669 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957688 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957703 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957780 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957808 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957829 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957849 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957865 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957880 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957910 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957928 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957944 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957958 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957974 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957988 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958003 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958017 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958032 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958047 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958061 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958076 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958090 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958110 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958125 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958142 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958156 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958172 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958185 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958203 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958218 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958233 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958248 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958264 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958278 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958291 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957246 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957353 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957318 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957537 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957662 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957689 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957843 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957870 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957924 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.957971 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958241 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958257 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.958304 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:17.458290478 +0000 UTC m=+18.867123864 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959272 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959282 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959291 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959310 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959329 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959357 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959376 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959391 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959407 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959421 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959436 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959450 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959464 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959478 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959492 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959506 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959537 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959553 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959569 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959585 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959587 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959601 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959617 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959636 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959651 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959668 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959673 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959684 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959741 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959763 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959780 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959795 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959811 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959823 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959829 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959892 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959914 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959933 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959949 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959965 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959982 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960000 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960016 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960031 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960049 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960065 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960079 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960094 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960115 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960130 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960146 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960161 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960177 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960192 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960208 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960225 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960240 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960256 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960274 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960294 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960312 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960328 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960356 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960371 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960387 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960404 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960422 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960439 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960456 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960471 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960485 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960500 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960514 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960548 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960570 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960585 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960601 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960618 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960634 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960649 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960663 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960679 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960700 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960715 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960730 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960747 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960767 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960782 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960798 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960813 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960830 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960847 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960864 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960880 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960896 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960912 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960946 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcc7l\" (UniqueName: \"kubernetes.io/projected/917c8c88-a0de-49af-9388-990df3c83b25-kube-api-access-hcc7l\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960966 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f24da5a0-0c94-4513-9c44-84818f240046-rootfs\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960989 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961007 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961024 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961044 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961059 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f24da5a0-0c94-4513-9c44-84818f240046-proxy-tls\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961076 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961093 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961108 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961128 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961143 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961160 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/917c8c88-a0de-49af-9388-990df3c83b25-hosts-file\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961177 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961193 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961208 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f24da5a0-0c94-4513-9c44-84818f240046-mcd-auth-proxy-config\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961222 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h25bh\" (UniqueName: \"kubernetes.io/projected/f24da5a0-0c94-4513-9c44-84818f240046-kube-api-access-h25bh\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961237 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961253 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961273 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961336 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961355 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961365 4708 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961375 4708 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961385 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961394 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961404 4708 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961414 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961423 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961432 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961443 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.962133 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967277 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968278 4708 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968581 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968943 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970840 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.974460 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975986 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d" exitCode=255 Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.976467 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d"} Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959845 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959854 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958361 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958469 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958492 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958497 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958507 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958633 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958655 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958668 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958726 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958773 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958793 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958841 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958893 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958970 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958984 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959002 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959053 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959126 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959168 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.959257 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960087 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960154 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960395 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960429 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960565 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960881 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.960901 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.961532 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963022 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963083 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963213 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963268 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963444 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963562 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963606 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963756 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.963823 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964180 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964187 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964245 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964437 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964462 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964475 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964644 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964742 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964811 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964872 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.964904 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965053 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965123 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965127 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965224 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965626 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.965661 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.966593 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.966806 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.981917 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.966835 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.966968 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967027 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967159 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967259 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967306 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967385 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967405 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967431 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.967494 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982095 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:17.482080522 +0000 UTC m=+18.890913907 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967576 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967714 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.967914 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968000 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.968169 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.958323 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968475 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968820 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.968939 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969222 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969255 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969275 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969393 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969506 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969533 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969542 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969764 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969776 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969784 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.969854 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970186 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970623 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970695 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970722 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970730 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970735 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.970897 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971000 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971021 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971199 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971249 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971421 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971485 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971656 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971917 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.971990 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.972151 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.972494 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.972697 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.972864 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973252 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973339 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973532 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973536 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973655 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973880 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973925 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.973976 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.974105 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.974946 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.975000 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982501 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982511 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975218 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975328 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975331 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975506 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975711 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.975908 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.976287 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.976729 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.976920 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.977193 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.978332 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.980014 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982636 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982667 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982201 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:17.482186359 +0000 UTC m=+18.891019744 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982893 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:17.482882028 +0000 UTC m=+18.891715414 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.982999 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:17.482985792 +0000 UTC m=+18.891819168 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.983987 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.984059 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.984479 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.987769 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.987883 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.988410 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.989389 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.989456 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.989544 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.989590 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990113 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990790 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990829 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990252 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990854 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.990131 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.991014 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.992970 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993182 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993285 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993303 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993317 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993486 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993701 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993718 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.993892 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.994061 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.994574 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995107 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995441 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995616 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995822 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.996047 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.996950 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.997869 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.998336 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.998092 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.998122 4708 scope.go:117] "RemoveContainer" containerID="713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995634 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:16 crc kubenswrapper[4708]: E1125 05:41:16.998190 4708 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.998804 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.999030 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.998209 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.995295 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:16 crc kubenswrapper[4708]: I1125 05:41:16.994704 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:16.995276 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.000792 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.001752 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.003797 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.005408 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.013064 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.017391 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.023278 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.031713 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.034375 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.036643 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.038593 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.039165 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.044632 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.050731 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.056795 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061850 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/917c8c88-a0de-49af-9388-990df3c83b25-hosts-file\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061887 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f24da5a0-0c94-4513-9c44-84818f240046-mcd-auth-proxy-config\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061905 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h25bh\" (UniqueName: \"kubernetes.io/projected/f24da5a0-0c94-4513-9c44-84818f240046-kube-api-access-h25bh\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061922 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061936 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061921 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/917c8c88-a0de-49af-9388-990df3c83b25-hosts-file\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061957 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f24da5a0-0c94-4513-9c44-84818f240046-rootfs\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062013 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062041 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/f24da5a0-0c94-4513-9c44-84818f240046-rootfs\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062052 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.061977 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcc7l\" (UniqueName: \"kubernetes.io/projected/917c8c88-a0de-49af-9388-990df3c83b25-kube-api-access-hcc7l\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062082 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f24da5a0-0c94-4513-9c44-84818f240046-proxy-tls\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062327 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062352 4708 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062362 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062370 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062379 4708 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062387 4708 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062394 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062410 4708 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062421 4708 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062429 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062436 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062445 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062453 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062460 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062468 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062475 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062483 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062490 4708 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062498 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062505 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062512 4708 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062535 4708 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062543 4708 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062550 4708 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062557 4708 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062565 4708 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062573 4708 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062580 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062587 4708 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062595 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062603 4708 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062611 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062619 4708 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062628 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062636 4708 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062645 4708 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062654 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062663 4708 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062671 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062679 4708 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062687 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062695 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062703 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062712 4708 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062720 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062728 4708 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062735 4708 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062743 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062752 4708 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062759 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062767 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062775 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062784 4708 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062791 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062799 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062807 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062815 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062823 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062831 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062838 4708 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062846 4708 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062853 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062861 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062868 4708 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062875 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062882 4708 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062890 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062897 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062905 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062912 4708 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062920 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062927 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062934 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062941 4708 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062949 4708 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062957 4708 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062965 4708 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062972 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062980 4708 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062988 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062996 4708 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063003 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063010 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063017 4708 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063025 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.062933 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063032 4708 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063723 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063735 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063745 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063754 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063764 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063774 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063781 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063789 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063797 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063805 4708 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063812 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063842 4708 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063850 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063858 4708 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063865 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063872 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063879 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063886 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063893 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063900 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063907 4708 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063914 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063922 4708 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063932 4708 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063940 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063948 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063955 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063962 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063970 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063977 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063985 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063993 4708 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064002 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064011 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064018 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064026 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064033 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064041 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064049 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064057 4708 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064064 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064072 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064079 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064086 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064094 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064101 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064108 4708 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064114 4708 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064122 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064129 4708 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064137 4708 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064144 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064151 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064158 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064165 4708 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064173 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064181 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064190 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064441 4708 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064455 4708 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064464 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.063143 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f24da5a0-0c94-4513-9c44-84818f240046-mcd-auth-proxy-config\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064497 4708 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064500 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f24da5a0-0c94-4513-9c44-84818f240046-proxy-tls\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064507 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064564 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064573 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064582 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064590 4708 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064598 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064607 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064615 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064629 4708 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064638 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064645 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064653 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064661 4708 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064668 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064675 4708 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064683 4708 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064690 4708 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064698 4708 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064705 4708 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064713 4708 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064720 4708 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064730 4708 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064737 4708 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064744 4708 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064752 4708 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064760 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064768 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064776 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064783 4708 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064791 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064798 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064805 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064813 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064821 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.064828 4708 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.073593 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.074531 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h25bh\" (UniqueName: \"kubernetes.io/projected/f24da5a0-0c94-4513-9c44-84818f240046-kube-api-access-h25bh\") pod \"machine-config-daemon-lh62q\" (UID: \"f24da5a0-0c94-4513-9c44-84818f240046\") " pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.077044 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcc7l\" (UniqueName: \"kubernetes.io/projected/917c8c88-a0de-49af-9388-990df3c83b25-kube-api-access-hcc7l\") pod \"node-resolver-clxd5\" (UID: \"917c8c88-a0de-49af-9388-990df3c83b25\") " pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.080938 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.088211 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.093336 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.165447 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.169568 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.176081 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 05:41:17 crc kubenswrapper[4708]: W1125 05:41:17.177652 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-b952e799cf5e8b5e4dff536b52d02e247759446742f818ff75cfbed0dbdb822e WatchSource:0}: Error finding container b952e799cf5e8b5e4dff536b52d02e247759446742f818ff75cfbed0dbdb822e: Status 404 returned error can't find the container with id b952e799cf5e8b5e4dff536b52d02e247759446742f818ff75cfbed0dbdb822e Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.180694 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-clxd5" Nov 25 05:41:17 crc kubenswrapper[4708]: W1125 05:41:17.189272 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-d2d954376bcbeda00ae254f4fe28eea26ce1a87a355f910a33bbc43dfece6c6f WatchSource:0}: Error finding container d2d954376bcbeda00ae254f4fe28eea26ce1a87a355f910a33bbc43dfece6c6f: Status 404 returned error can't find the container with id d2d954376bcbeda00ae254f4fe28eea26ce1a87a355f910a33bbc43dfece6c6f Nov 25 05:41:17 crc kubenswrapper[4708]: W1125 05:41:17.196015 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod917c8c88_a0de_49af_9388_990df3c83b25.slice/crio-507d49fccb1e759b3deb5ef01d899878a4bc374ab09a06eafaca52e0a651b878 WatchSource:0}: Error finding container 507d49fccb1e759b3deb5ef01d899878a4bc374ab09a06eafaca52e0a651b878: Status 404 returned error can't find the container with id 507d49fccb1e759b3deb5ef01d899878a4bc374ab09a06eafaca52e0a651b878 Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.216638 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.268730 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-rx7k7"] Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.269195 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-rjrnw"] Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.269304 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qppcd"] Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.269347 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.269408 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.270038 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283048 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283600 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283609 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283781 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283850 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283899 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283945 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.283984 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.284028 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.284055 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.284112 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.284200 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.284313 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.292773 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.325798 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.350747 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.369913 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.369977 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7wnr\" (UniqueName: \"kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370000 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-netns\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370018 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370039 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-socket-dir-parent\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370069 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370086 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370101 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370119 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-k8s-cni-cncf-io\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370137 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-conf-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370154 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cnibin\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370168 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370182 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370201 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370218 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5fqd\" (UniqueName: \"kubernetes.io/projected/abcfd1aa-ec01-4cae-8858-0d7b140c6344-kube-api-access-w5fqd\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370242 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370258 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cnibin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370292 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-etc-kubernetes\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370349 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370365 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-os-release\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370382 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-multus\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370399 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-kubelet\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370415 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-daemon-config\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370431 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw4xn\" (UniqueName: \"kubernetes.io/projected/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-kube-api-access-mw4xn\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370446 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370461 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370480 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-bin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370500 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370748 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370777 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cni-binary-copy\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370884 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-hostroot\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370946 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370972 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.370988 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371008 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371021 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371035 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371048 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371087 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-multus-certs\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371106 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-os-release\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371124 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371137 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371155 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-system-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.371170 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-system-cni-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.388041 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.403331 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.414332 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.424428 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.432353 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.440849 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.450147 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.457030 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.463448 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.471172 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472115 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472212 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-system-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.472294 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:18.4722723 +0000 UTC m=+19.881105686 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472320 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-system-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472331 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-system-cni-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472384 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472405 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7wnr\" (UniqueName: \"kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472423 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-netns\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472441 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472461 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472472 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-system-cni-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472480 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472500 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-socket-dir-parent\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472504 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-netns\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472545 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472564 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472581 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-k8s-cni-cncf-io\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-conf-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472617 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cnibin\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472659 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472673 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5fqd\" (UniqueName: \"kubernetes.io/projected/abcfd1aa-ec01-4cae-8858-0d7b140c6344-kube-api-access-w5fqd\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472703 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-etc-kubernetes\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472727 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472740 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cnibin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472756 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472766 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-conf-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472770 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-daemon-config\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472802 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-os-release\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472831 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-multus\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472849 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-kubelet\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472869 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472889 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw4xn\" (UniqueName: \"kubernetes.io/projected/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-kube-api-access-mw4xn\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472906 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472924 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472939 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-bin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472963 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472977 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cni-binary-copy\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.472991 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-hostroot\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473011 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473030 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473045 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473059 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473075 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473089 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473103 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473144 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-multus-certs\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473160 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-os-release\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473202 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473217 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473258 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473375 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-daemon-config\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473421 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cnibin\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473823 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473868 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-kubelet\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473893 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473930 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-binary-copy\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473974 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-socket-dir-parent\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.473999 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474020 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474148 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-k8s-cni-cncf-io\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474152 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-bin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474213 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-etc-kubernetes\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474311 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474372 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-multus-cni-dir\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474416 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474446 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474471 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474481 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-os-release\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474501 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-hostroot\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474489 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cni-binary-copy\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474512 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474560 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-cnibin\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474557 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474589 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474595 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-run-multus-certs\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474567 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474620 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-os-release\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474600 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474664 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/abcfd1aa-ec01-4cae-8858-0d7b140c6344-host-var-lib-cni-multus\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474678 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474695 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.474756 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.475468 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-tuning-conf-dir\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.476705 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.481841 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.486368 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw4xn\" (UniqueName: \"kubernetes.io/projected/e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d-kube-api-access-mw4xn\") pod \"multus-additional-cni-plugins-rx7k7\" (UID: \"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\") " pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.486671 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7wnr\" (UniqueName: \"kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr\") pod \"ovnkube-node-qppcd\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.488383 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5fqd\" (UniqueName: \"kubernetes.io/projected/abcfd1aa-ec01-4cae-8858-0d7b140c6344-kube-api-access-w5fqd\") pod \"multus-rjrnw\" (UID: \"abcfd1aa-ec01-4cae-8858-0d7b140c6344\") " pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.499131 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.538579 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.574011 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.574140 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.574239 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.574343 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574166 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574557 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:18.574539067 +0000 UTC m=+19.983372453 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574265 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574702 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574324 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574818 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574843 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574465 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574764 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574925 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:18.574897306 +0000 UTC m=+19.983730692 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574946 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:18.574938072 +0000 UTC m=+19.983771458 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.574958 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:18.57495277 +0000 UTC m=+19.983786156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.580127 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.599702 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.613141 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-rjrnw" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.618084 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.623467 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.659038 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 05:41:17 crc kubenswrapper[4708]: W1125 05:41:17.677862 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4dfd5bc_2e9a_4b1c_a99b_526d5c58be1d.slice/crio-093937b1b5df3df6c34816a47173bdb41d6414e1911d7a96fe81b5cf60e56596 WatchSource:0}: Error finding container 093937b1b5df3df6c34816a47173bdb41d6414e1911d7a96fe81b5cf60e56596: Status 404 returned error can't find the container with id 093937b1b5df3df6c34816a47173bdb41d6414e1911d7a96fe81b5cf60e56596 Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.702576 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.744438 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.781546 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.822863 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.862496 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.893205 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:17 crc kubenswrapper[4708]: E1125 05:41:17.893360 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.902129 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.942429 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:17Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.988548 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.988605 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.988618 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b952e799cf5e8b5e4dff536b52d02e247759446742f818ff75cfbed0dbdb822e"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.990795 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.992583 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.992786 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.994243 4708 generic.go:334] "Generic (PLEG): container finished" podID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerID="b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b" exitCode=0 Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.994319 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.994393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"2ac6445092c4cdcbc3bd15dacca5f99b8cc64c35e81d1c36d26b4e7179ff3ff7"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.995342 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerStarted","Data":"093937b1b5df3df6c34816a47173bdb41d6414e1911d7a96fe81b5cf60e56596"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.997120 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.997155 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.997168 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"a4385b4e4e57c9c445a64a5faa3ecb2d9537d103bb3446beb54561f63bddbea5"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.998315 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd"} Nov 25 05:41:17 crc kubenswrapper[4708]: I1125 05:41:17.998354 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6b01960d6c522c4cb8d4935b6b78c1246864701264e8ff82940fea43d8ed0f22"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.003177 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d2d954376bcbeda00ae254f4fe28eea26ce1a87a355f910a33bbc43dfece6c6f"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.004663 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerStarted","Data":"cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.004734 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerStarted","Data":"d90b21d3847ef9a2eafd4204c2c9fa910f7e7ef7aec438dfd787dfdaa935bc87"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.006580 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-clxd5" event={"ID":"917c8c88-a0de-49af-9388-990df3c83b25","Type":"ContainerStarted","Data":"95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.006628 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-clxd5" event={"ID":"917c8c88-a0de-49af-9388-990df3c83b25","Type":"ContainerStarted","Data":"507d49fccb1e759b3deb5ef01d899878a4bc374ab09a06eafaca52e0a651b878"} Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.007410 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.021763 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.060266 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.102682 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.143673 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.182122 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.221139 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.261983 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.302222 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.345907 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.385809 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.420424 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.460595 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.486823 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.487037 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:20.487009214 +0000 UTC m=+21.895842601 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.502681 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.540693 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.587575 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.587631 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.587665 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.587689 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587754 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587779 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587777 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587793 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587809 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587834 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587841 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:20.587825854 +0000 UTC m=+21.996659239 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587784 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587883 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:20.587877389 +0000 UTC m=+21.996710775 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587846 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587895 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:20.587889422 +0000 UTC m=+21.996722808 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.587934 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:20.587918837 +0000 UTC m=+21.996752223 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.590220 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.640239 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.662854 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.700983 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.742497 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.781744 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.824325 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.860696 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.893137 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.893151 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.893262 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:18 crc kubenswrapper[4708]: E1125 05:41:18.893346 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.899215 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.900328 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.901577 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.902203 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.903232 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.903828 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.904445 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.905428 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.906095 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.906564 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.907498 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.908327 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.909643 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.910169 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.910717 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.911745 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.912251 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.913217 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.913660 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.914197 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.915258 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.915834 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.916793 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.917218 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.918278 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.918762 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.919347 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.920400 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.920882 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.921882 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.922378 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.923216 4708 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.923330 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.925016 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.925905 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.926346 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.927851 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.928465 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.929305 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.929924 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.931004 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.931458 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.932373 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.932991 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.933901 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.934344 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.935187 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.935715 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.936710 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.937166 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.937980 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.938424 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.939251 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.939899 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.940334 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.942994 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:18 crc kubenswrapper[4708]: I1125 05:41:18.980831 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.016887 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.017004 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.017065 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.017119 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.017172 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.017233 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.018472 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc" exitCode=0 Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.018602 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.023934 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.058611 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.100898 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-5wgns"] Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.101491 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.101988 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.114365 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.134577 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.154873 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.174779 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.194210 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5811ead-bd59-4da5-a686-bc1c974f86eb-serviceca\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.195006 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5811ead-bd59-4da5-a686-bc1c974f86eb-host\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.195044 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fqm8\" (UniqueName: \"kubernetes.io/projected/b5811ead-bd59-4da5-a686-bc1c974f86eb-kube-api-access-9fqm8\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.222041 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.261250 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.296379 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5811ead-bd59-4da5-a686-bc1c974f86eb-serviceca\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.296427 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5811ead-bd59-4da5-a686-bc1c974f86eb-host\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.296453 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fqm8\" (UniqueName: \"kubernetes.io/projected/b5811ead-bd59-4da5-a686-bc1c974f86eb-kube-api-access-9fqm8\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.296570 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b5811ead-bd59-4da5-a686-bc1c974f86eb-host\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.297447 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/b5811ead-bd59-4da5-a686-bc1c974f86eb-serviceca\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.301591 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.332937 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fqm8\" (UniqueName: \"kubernetes.io/projected/b5811ead-bd59-4da5-a686-bc1c974f86eb-kube-api-access-9fqm8\") pod \"node-ca-5wgns\" (UID: \"b5811ead-bd59-4da5-a686-bc1c974f86eb\") " pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.361222 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.402169 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.414601 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-5wgns" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.441993 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: W1125 05:41:19.444884 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5811ead_bd59_4da5_a686_bc1c974f86eb.slice/crio-12829b8062cc1104806bf706160cbd4b2baf941254c20de91b57f48b07d225e8 WatchSource:0}: Error finding container 12829b8062cc1104806bf706160cbd4b2baf941254c20de91b57f48b07d225e8: Status 404 returned error can't find the container with id 12829b8062cc1104806bf706160cbd4b2baf941254c20de91b57f48b07d225e8 Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.455301 4708 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.456995 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.457024 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.457034 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.457132 4708 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.506386 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.515280 4708 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.515427 4708 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.516727 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.516760 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.516772 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.516794 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.516806 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.530568 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.533058 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.533112 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.533124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.533142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.533392 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.543748 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.546704 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.546786 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.546801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.546843 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.546855 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.557164 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.559906 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.559938 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.559948 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.559962 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.559975 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.560978 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.568155 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.570692 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.570722 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.570731 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.570745 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.570755 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.581198 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.581324 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.582612 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.582642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.582652 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.582667 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.582679 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.601107 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.641424 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.680277 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.684967 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.685008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.685020 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.685038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.685049 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.721358 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.761994 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.787153 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.787186 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.787197 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.787212 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.787225 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.801438 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.840906 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.883663 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.889566 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.889611 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.889622 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.889638 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.889650 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.892879 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:19 crc kubenswrapper[4708]: E1125 05:41:19.893024 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.921491 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.962835 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.991276 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.991327 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.991348 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.991368 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:19 crc kubenswrapper[4708]: I1125 05:41:19.991382 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:19Z","lastTransitionTime":"2025-11-25T05:41:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.001537 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.022349 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5wgns" event={"ID":"b5811ead-bd59-4da5-a686-bc1c974f86eb","Type":"ContainerStarted","Data":"0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.022400 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-5wgns" event={"ID":"b5811ead-bd59-4da5-a686-bc1c974f86eb","Type":"ContainerStarted","Data":"12829b8062cc1104806bf706160cbd4b2baf941254c20de91b57f48b07d225e8"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.024336 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846" exitCode=0 Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.024377 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.025693 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.043289 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.090362 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.093997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.094041 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.094053 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.094068 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.094079 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.120362 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.161014 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.196495 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.196554 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.196565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.196584 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.196597 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.202168 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.240767 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.280333 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.298920 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.298950 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.298960 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.298978 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.298989 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.320631 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.361421 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.400353 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.400376 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.400386 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.400429 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.400439 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.405031 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.441081 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.480631 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.502705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.502732 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.502741 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.502756 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.502764 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.507050 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.507192 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:24.507176105 +0000 UTC m=+25.916009491 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.520699 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.559024 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.600473 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.604918 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.605000 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.605073 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.605129 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.605184 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.608283 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.608325 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.608347 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.608372 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608422 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608448 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608469 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608482 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608448 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608536 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608473 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:24.608460633 +0000 UTC m=+26.017294019 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608544 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608566 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:24.608555972 +0000 UTC m=+26.017389358 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608578 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:24.608573003 +0000 UTC m=+26.017406390 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608575 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.608687 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:24.608660307 +0000 UTC m=+26.017493693 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.640712 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.681244 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.706821 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.706935 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.706997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.707057 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.707107 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.723530 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.758781 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:20Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.809439 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.809480 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.809494 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.809514 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.809542 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.892482 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.892558 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.892629 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:20 crc kubenswrapper[4708]: E1125 05:41:20.892682 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.911410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.911458 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.911472 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.911489 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:20 crc kubenswrapper[4708]: I1125 05:41:20.911500 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:20Z","lastTransitionTime":"2025-11-25T05:41:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.014112 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.014180 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.014194 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.014212 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.014227 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.034620 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.036760 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252" exitCode=0 Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.036901 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.046958 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.060441 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.075074 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.085232 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.094268 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.102853 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.110043 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.116001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.116082 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.116132 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.116209 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.116270 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.119714 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.130163 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.162016 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.201674 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.218354 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.218402 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.218412 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.218430 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.218443 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.241100 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.280262 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.319207 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:21Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.320670 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.320729 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.320742 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.320757 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.320773 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.422491 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.422550 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.422560 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.422574 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.422586 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.524383 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.524435 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.524449 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.524464 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.524475 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.625941 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.625974 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.625985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.625997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.626007 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.727940 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.727977 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.727986 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.727999 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.728008 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.829467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.829499 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.829510 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.829535 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.829544 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.892432 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:21 crc kubenswrapper[4708]: E1125 05:41:21.892556 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.930955 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.930985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.930996 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.931008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:21 crc kubenswrapper[4708]: I1125 05:41:21.931018 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:21Z","lastTransitionTime":"2025-11-25T05:41:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.032227 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.032265 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.032286 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.032301 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.032313 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.044144 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d" exitCode=0 Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.044183 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.053972 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.061767 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.069878 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.077767 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.084543 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.096745 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.105041 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.112691 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.119505 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.128282 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.134889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.135210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.135221 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.135234 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.135243 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.136822 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.145072 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.154327 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.162354 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:22Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.237713 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.237750 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.237760 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.237776 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.237787 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.339506 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.339551 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.339562 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.339573 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.339580 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.441256 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.441294 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.441303 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.441313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.441320 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.542975 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.543009 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.543018 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.543032 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.543040 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.645358 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.645402 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.645413 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.645431 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.645442 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.747603 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.747633 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.747642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.747654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.747663 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.849982 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.850014 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.850024 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.850039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.850047 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.892692 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.892700 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:22 crc kubenswrapper[4708]: E1125 05:41:22.892832 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:22 crc kubenswrapper[4708]: E1125 05:41:22.892887 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.951772 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.951888 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.951897 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.951909 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:22 crc kubenswrapper[4708]: I1125 05:41:22.951920 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:22Z","lastTransitionTime":"2025-11-25T05:41:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.050733 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.050950 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053113 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053161 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053174 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053511 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58" exitCode=0 Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.053554 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.064044 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.072553 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.075312 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.082470 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.093331 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.103773 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.113949 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.123613 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160698 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160741 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160787 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160804 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.160815 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.175364 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.188333 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.204039 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.215125 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.224689 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.232973 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.243371 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.254207 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.262946 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.263042 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.263118 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.263200 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.263269 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.264011 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.271412 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.280662 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.288968 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.297381 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.306303 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.313363 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.322064 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.330542 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.338138 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.347686 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.362141 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:23Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.365533 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.365559 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.365569 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.365585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.365597 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.467730 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.467762 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.467771 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.467784 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.467797 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.569761 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.569789 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.569798 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.569809 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.569821 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.672024 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.672063 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.672071 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.672086 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.672095 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.773997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.774023 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.774031 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.774041 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.774049 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.875232 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.875278 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.875287 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.875300 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.875313 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.892569 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:23 crc kubenswrapper[4708]: E1125 05:41:23.892684 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.976625 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.976648 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.976656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.976681 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:23 crc kubenswrapper[4708]: I1125 05:41:23.976690 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:23Z","lastTransitionTime":"2025-11-25T05:41:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.068164 4708 generic.go:334] "Generic (PLEG): container finished" podID="e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d" containerID="c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c" exitCode=0 Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.068805 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerDied","Data":"c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.068837 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.068849 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.078211 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.078234 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.078243 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.078262 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.078272 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.079909 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.094415 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.097746 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.111990 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.124900 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.132591 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.143716 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.153073 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.161882 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.171777 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.179682 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.179703 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.179712 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.179723 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.179733 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.186331 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.195761 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.207681 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.216962 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.224101 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.239690 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.250159 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.259147 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.267401 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.277058 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.282198 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.282298 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.282366 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.282461 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.282541 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.284614 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.292481 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.301624 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.309577 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.317875 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.325927 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.334574 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.343618 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.350593 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:24Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.385201 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.385232 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.385242 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.385268 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.385280 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.487742 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.487790 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.487802 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.487818 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.487830 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.552799 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.552954 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:32.552930769 +0000 UTC m=+33.961764156 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.590234 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.590269 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.590278 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.590290 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.590298 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.654440 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.654487 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.654509 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.654547 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.654643 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.654657 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.654667 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.654705 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:32.654693662 +0000 UTC m=+34.063527049 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655301 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655319 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655329 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655351 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:32.655344271 +0000 UTC m=+34.064177657 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655378 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655397 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:32.65539213 +0000 UTC m=+34.064225516 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655429 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.655447 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:32.655442495 +0000 UTC m=+34.064275881 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.692987 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.693026 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.693038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.693057 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.693071 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.794960 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.794994 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.795002 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.795015 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.795028 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.892998 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.892998 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.893126 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:24 crc kubenswrapper[4708]: E1125 05:41:24.893201 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.896569 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.896596 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.896605 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.896614 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.896621 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.999358 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.999729 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.999740 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.999755 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:24 crc kubenswrapper[4708]: I1125 05:41:24.999765 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:24Z","lastTransitionTime":"2025-11-25T05:41:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.074998 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/0.log" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.077809 4708 generic.go:334] "Generic (PLEG): container finished" podID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerID="9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5" exitCode=1 Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.077897 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.078435 4708 scope.go:117] "RemoveContainer" containerID="9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.085337 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" event={"ID":"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d","Type":"ContainerStarted","Data":"7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.093163 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.101785 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.101842 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.101853 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.101874 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.101891 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.103726 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.114159 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.125655 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.139825 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"message\\\":\\\"05:41:24.975322 5995 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 05:41:24.975339 5995 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 05:41:24.975429 5995 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 05:41:24.975434 5995 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 05:41:24.975441 5995 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 05:41:24.975490 5995 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 05:41:24.975516 5995 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 05:41:24.975487 5995 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 05:41:24.975591 5995 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 05:41:24.975617 5995 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 05:41:24.975632 5995 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 05:41:24.975625 5995 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 05:41:24.975679 5995 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 05:41:24.975709 5995 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 05:41:24.975778 5995 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 05:41:24.975780 5995 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.151867 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.165211 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.176301 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.184939 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.194035 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.202667 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.204443 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.204470 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.204480 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.204497 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.204508 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.212984 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.222642 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.230870 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.241302 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.250152 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.258098 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.269018 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.281622 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"message\\\":\\\"05:41:24.975322 5995 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 05:41:24.975339 5995 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 05:41:24.975429 5995 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 05:41:24.975434 5995 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 05:41:24.975441 5995 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 05:41:24.975490 5995 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 05:41:24.975516 5995 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 05:41:24.975487 5995 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 05:41:24.975591 5995 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 05:41:24.975617 5995 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 05:41:24.975632 5995 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 05:41:24.975625 5995 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 05:41:24.975679 5995 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 05:41:24.975709 5995 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 05:41:24.975778 5995 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 05:41:24.975780 5995 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.290823 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.303621 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.307111 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.307142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.307155 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.307177 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.307189 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.313360 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.322289 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.337012 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.347168 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.375138 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.390739 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.406054 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.409442 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.409488 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.409500 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.409540 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.409552 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.511487 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.511545 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.511557 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.511571 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.511581 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.613961 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.614010 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.614019 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.614038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.614048 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.716336 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.716379 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.716390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.716407 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.716419 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.818846 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.818884 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.818896 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.818910 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.818920 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.892938 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:25 crc kubenswrapper[4708]: E1125 05:41:25.893052 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.921069 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.921105 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.921114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.921130 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:25 crc kubenswrapper[4708]: I1125 05:41:25.921147 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:25Z","lastTransitionTime":"2025-11-25T05:41:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.023685 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.023746 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.023759 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.023783 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.023798 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.090497 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/1.log" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.091259 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/0.log" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.094306 4708 generic.go:334] "Generic (PLEG): container finished" podID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerID="8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827" exitCode=1 Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.094346 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.094421 4708 scope.go:117] "RemoveContainer" containerID="9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.095003 4708 scope.go:117] "RemoveContainer" containerID="8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827" Nov 25 05:41:26 crc kubenswrapper[4708]: E1125 05:41:26.095282 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.106505 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.116508 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.126087 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.126118 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.126128 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.126144 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.126156 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.127224 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.141983 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.149823 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.160975 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.174427 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e221e57c4836cf7e333c0729e8fd9348c8d6711cd9baad11b0dc3e2bfd784b5\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"message\\\":\\\"05:41:24.975322 5995 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 05:41:24.975339 5995 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 05:41:24.975429 5995 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1125 05:41:24.975434 5995 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 05:41:24.975441 5995 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1125 05:41:24.975490 5995 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 05:41:24.975516 5995 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 05:41:24.975487 5995 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1125 05:41:24.975591 5995 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1125 05:41:24.975617 5995 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 05:41:24.975632 5995 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 05:41:24.975625 5995 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 05:41:24.975679 5995 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 05:41:24.975709 5995 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 05:41:24.975778 5995 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 05:41:24.975780 5995 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.185405 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.194691 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.202977 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.217009 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.224591 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.228454 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.228493 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.228506 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.228541 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.228555 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.233196 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.241743 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:26Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.330866 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.330917 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.330928 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.330946 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.330960 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.433210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.433266 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.433277 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.433295 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.433309 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.535543 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.535579 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.535589 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.535605 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.535619 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.637490 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.637556 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.637568 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.637585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.637596 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.739956 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.739991 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.739999 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.740012 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.740022 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.841839 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.841880 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.841890 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.841924 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.841937 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.892832 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.892893 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:26 crc kubenswrapper[4708]: E1125 05:41:26.892988 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:26 crc kubenswrapper[4708]: E1125 05:41:26.893214 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.944165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.944204 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.944216 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.944253 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:26 crc kubenswrapper[4708]: I1125 05:41:26.944270 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:26Z","lastTransitionTime":"2025-11-25T05:41:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.047022 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.047066 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.047078 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.047095 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.047106 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.098289 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/1.log" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.101305 4708 scope.go:117] "RemoveContainer" containerID="8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827" Nov 25 05:41:27 crc kubenswrapper[4708]: E1125 05:41:27.101451 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.113535 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.124080 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.132592 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.143668 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.149062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.149091 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.149100 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.149112 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.149122 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.150821 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.158651 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.167603 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.179688 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.188539 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.196931 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.206757 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.213003 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.221688 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.230253 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:27Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.251857 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.251897 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.251911 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.251933 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.251944 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.354388 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.354424 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.354433 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.354451 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.354461 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.456565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.456603 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.456630 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.456644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.456655 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.558823 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.558856 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.558865 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.558880 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.558890 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.660860 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.660916 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.660927 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.660939 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.660951 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.762688 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.762798 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.762827 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.762843 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.762854 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.864672 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.864700 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.864709 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.864723 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.864731 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.892317 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:27 crc kubenswrapper[4708]: E1125 05:41:27.892438 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.966752 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.966779 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.966788 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.966799 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:27 crc kubenswrapper[4708]: I1125 05:41:27.966807 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:27Z","lastTransitionTime":"2025-11-25T05:41:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.068696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.068745 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.068755 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.068771 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.068783 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.170944 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.170983 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.170993 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.171008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.171018 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.272590 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.272621 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.272631 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.272644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.272672 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.374576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.374608 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.374639 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.374652 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.374661 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.476547 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.476577 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.476588 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.476599 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.476607 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.578263 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.578298 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.578307 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.578320 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.578329 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.680626 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.680676 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.680688 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.680704 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.680718 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.782827 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.782882 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.782893 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.782905 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.782914 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.884505 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.884562 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.884576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.884592 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.884603 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.893255 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.893309 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:28 crc kubenswrapper[4708]: E1125 05:41:28.893421 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:28 crc kubenswrapper[4708]: E1125 05:41:28.893587 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.906980 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.917252 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.930692 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.940425 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.952011 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.961477 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.968398 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.977379 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.986923 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.986960 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.986970 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.986983 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.986993 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:28Z","lastTransitionTime":"2025-11-25T05:41:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.989709 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:28 crc kubenswrapper[4708]: I1125 05:41:28.999015 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:28Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.007996 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.015751 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.025599 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.033744 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.089338 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.089376 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.089390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.089408 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.089424 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.191431 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.191467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.191476 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.191490 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.191502 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.293729 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.293768 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.293777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.293792 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.293805 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.396406 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.396458 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.396468 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.396485 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.396497 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.462790 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht"] Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.463338 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.465169 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.465246 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.474437 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.486476 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.498875 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.498917 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.498929 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.498947 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.498960 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.499982 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.514224 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.522715 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.531385 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.540659 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.549060 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.556180 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.564931 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.573176 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.581686 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.590190 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.598385 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.600504 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swwd7\" (UniqueName: \"kubernetes.io/projected/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-kube-api-access-swwd7\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.600562 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.600610 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.600632 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.600954 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.601004 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.601015 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.601032 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.601045 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.605819 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.701264 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.701300 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.701341 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swwd7\" (UniqueName: \"kubernetes.io/projected/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-kube-api-access-swwd7\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.701369 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702103 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702202 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-env-overrides\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702615 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702623 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702634 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.702643 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.707393 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.714315 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swwd7\" (UniqueName: \"kubernetes.io/projected/a1da4d69-b614-43c1-ab20-37ac4b7e1e11-kube-api-access-swwd7\") pod \"ovnkube-control-plane-749d76644c-nq7ht\" (UID: \"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.774345 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" Nov 25 05:41:29 crc kubenswrapper[4708]: W1125 05:41:29.788533 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1da4d69_b614_43c1_ab20_37ac4b7e1e11.slice/crio-cfb30479db69f8a74244731e2e6cb69b1368e236b095345975a02e57f255cb02 WatchSource:0}: Error finding container cfb30479db69f8a74244731e2e6cb69b1368e236b095345975a02e57f255cb02: Status 404 returned error can't find the container with id cfb30479db69f8a74244731e2e6cb69b1368e236b095345975a02e57f255cb02 Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.808150 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.808198 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.808211 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.808237 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.808252 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.892888 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.893042 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.896385 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.896416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.896426 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.896441 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.896453 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.909585 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.912805 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.912829 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.912838 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.912850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.912860 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.921983 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.926441 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.926558 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.926629 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.926694 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.926760 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.935837 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.939103 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.939215 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.939277 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.939338 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.939389 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.949642 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.952943 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.953035 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.953097 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.953161 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.953234 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.962641 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:29Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:29 crc kubenswrapper[4708]: E1125 05:41:29.962964 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.965309 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.965353 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.965362 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.965374 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:29 crc kubenswrapper[4708]: I1125 05:41:29.965383 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:29Z","lastTransitionTime":"2025-11-25T05:41:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.068139 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.068178 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.068195 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.068215 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.068227 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.109032 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" event={"ID":"a1da4d69-b614-43c1-ab20-37ac4b7e1e11","Type":"ContainerStarted","Data":"b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.109083 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" event={"ID":"a1da4d69-b614-43c1-ab20-37ac4b7e1e11","Type":"ContainerStarted","Data":"f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.109099 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" event={"ID":"a1da4d69-b614-43c1-ab20-37ac4b7e1e11","Type":"ContainerStarted","Data":"cfb30479db69f8a74244731e2e6cb69b1368e236b095345975a02e57f255cb02"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.122733 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.131661 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.140639 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.149038 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.156134 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.164077 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.170170 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.170202 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.170211 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.170224 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.170232 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.171073 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.179433 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-g6lks"] Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.179872 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.179920 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.180066 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.193101 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.201119 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.210768 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.219099 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.226464 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.233480 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.241135 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.247563 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.256669 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.263685 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.272625 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.272656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.272665 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.272679 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.272687 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.276179 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.288243 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.295437 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.302075 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.308095 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9fxm\" (UniqueName: \"kubernetes.io/projected/711da2a4-b5f2-40ce-84d4-619c4e3f2522-kube-api-access-h9fxm\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.308121 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.310467 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.318860 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.326689 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.333017 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.340699 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.348146 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.356211 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.364341 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.371558 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:30Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.374940 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.374984 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.374993 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.375008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.375016 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.409411 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9fxm\" (UniqueName: \"kubernetes.io/projected/711da2a4-b5f2-40ce-84d4-619c4e3f2522-kube-api-access-h9fxm\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.409442 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.409554 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.409609 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:30.909593306 +0000 UTC m=+32.318426692 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.421996 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9fxm\" (UniqueName: \"kubernetes.io/projected/711da2a4-b5f2-40ce-84d4-619c4e3f2522-kube-api-access-h9fxm\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.476965 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.476987 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.476997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.477027 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.477036 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.579416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.579448 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.579458 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.579473 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.579482 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.681159 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.681200 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.681210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.681225 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.681236 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.782801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.782830 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.782840 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.782850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.782859 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.884388 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.884422 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.884432 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.884449 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.884459 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.892664 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.892775 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.892670 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.892864 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.913495 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.913656 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:30 crc kubenswrapper[4708]: E1125 05:41:30.913711 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:31.91369683 +0000 UTC m=+33.322530216 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.986251 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.986283 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.986292 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.986304 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:30 crc kubenswrapper[4708]: I1125 05:41:30.986313 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:30Z","lastTransitionTime":"2025-11-25T05:41:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.087905 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.087941 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.087951 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.087966 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.087975 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.189616 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.189651 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.189660 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.189672 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.189683 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.291567 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.291601 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.291609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.291621 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.291630 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.393480 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.393511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.393537 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.393549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.393558 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.495618 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.495651 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.495662 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.495674 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.495682 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.597747 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.597783 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.597793 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.597808 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.597818 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.699640 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.699673 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.699682 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.699692 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.699700 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.802263 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.802283 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.802294 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.802305 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.802312 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.892513 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.892609 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:31 crc kubenswrapper[4708]: E1125 05:41:31.892715 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:31 crc kubenswrapper[4708]: E1125 05:41:31.892769 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.903938 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.903967 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.903976 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.903990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.904002 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:31Z","lastTransitionTime":"2025-11-25T05:41:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:31 crc kubenswrapper[4708]: I1125 05:41:31.923386 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:31 crc kubenswrapper[4708]: E1125 05:41:31.923491 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:31 crc kubenswrapper[4708]: E1125 05:41:31.923558 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:33.923545694 +0000 UTC m=+35.332379080 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.007366 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.007395 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.007405 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.007419 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.007428 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.109063 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.109121 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.109140 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.109160 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.109194 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.211609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.211644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.211654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.211665 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.211673 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.313472 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.313496 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.313505 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.313515 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.313543 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.416013 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.416066 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.416078 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.416090 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.416098 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.518315 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.518361 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.518386 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.518405 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.518421 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.620250 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.620287 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.620297 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.620312 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.620322 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.628560 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.628714 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:41:48.628696474 +0000 UTC m=+50.037529860 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.722310 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.722361 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.722388 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.722407 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.722421 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.729579 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.729623 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.729651 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.729675 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729750 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729829 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729893 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:48.72986246 +0000 UTC m=+50.138695856 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729937 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:48.72992141 +0000 UTC m=+50.138754807 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729846 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729982 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.730007 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.729786 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.730047 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.730051 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:48.730042827 +0000 UTC m=+50.138876223 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.730060 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.730114 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:48.730099473 +0000 UTC m=+50.138932859 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.824360 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.824625 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.824634 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.824643 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.824652 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.892661 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.892733 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.892845 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:32 crc kubenswrapper[4708]: E1125 05:41:32.892966 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.926311 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.926415 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.926492 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.926602 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:32 crc kubenswrapper[4708]: I1125 05:41:32.926678 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:32Z","lastTransitionTime":"2025-11-25T05:41:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.027977 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.027999 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.028009 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.028021 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.028030 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.130203 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.130236 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.130246 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.130260 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.130268 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.232366 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.232406 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.232416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.232436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.232449 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.334846 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.334879 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.334891 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.334907 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.334916 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.436841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.436877 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.436888 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.436901 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.436911 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.538990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.539025 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.539035 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.539048 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.539058 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.641022 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.641119 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.641202 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.641272 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.641336 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.742797 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.742829 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.742837 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.742851 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.742860 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.844638 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.844673 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.844685 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.844696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.844705 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.892653 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:33 crc kubenswrapper[4708]: E1125 05:41:33.892854 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.892684 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:33 crc kubenswrapper[4708]: E1125 05:41:33.893024 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.937329 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.939593 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:33 crc kubenswrapper[4708]: E1125 05:41:33.939698 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:33 crc kubenswrapper[4708]: E1125 05:41:33.939748 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:37.939736865 +0000 UTC m=+39.348570251 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.946469 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.946502 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.946511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.946549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.946559 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:33Z","lastTransitionTime":"2025-11-25T05:41:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.947719 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.957108 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.965649 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.973493 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.982904 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:33 crc kubenswrapper[4708]: I1125 05:41:33.994585 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:33Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.005362 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.013033 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.021345 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.029886 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.038362 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.044953 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.048729 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.048825 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.048883 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.048962 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.049037 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.052266 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.059737 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.067732 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.075410 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:34Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.150971 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.151010 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.151024 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.151044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.151057 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.252780 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.252927 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.253002 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.253074 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.253134 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.355126 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.355180 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.355191 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.355208 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.355217 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.457561 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.457691 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.457745 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.457797 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.457845 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.559775 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.559813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.559824 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.559839 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.559852 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.661693 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.661729 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.661742 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.661754 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.661764 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.764172 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.764207 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.764217 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.764228 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.764258 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.866051 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.866080 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.866090 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.866103 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.866112 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.892886 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:34 crc kubenswrapper[4708]: E1125 05:41:34.892981 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.893189 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:34 crc kubenswrapper[4708]: E1125 05:41:34.893320 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.967853 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.967893 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.967906 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.967920 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:34 crc kubenswrapper[4708]: I1125 05:41:34.967930 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:34Z","lastTransitionTime":"2025-11-25T05:41:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.069386 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.069435 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.069447 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.069459 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.069467 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.170688 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.170832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.170898 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.170963 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.171027 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.272817 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.272848 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.272858 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.272889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.272898 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.374254 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.374285 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.374294 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.374312 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.374324 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.475997 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.476023 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.476032 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.476043 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.476051 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.577962 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.578006 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.578020 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.578037 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.578051 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.679982 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.680017 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.680031 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.680047 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.680057 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.781271 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.781620 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.781690 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.781758 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.781825 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.883760 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.883807 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.883819 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.883831 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.883839 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.893224 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.893230 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:35 crc kubenswrapper[4708]: E1125 05:41:35.893405 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:35 crc kubenswrapper[4708]: E1125 05:41:35.893561 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.985497 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.985549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.985566 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.985582 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:35 crc kubenswrapper[4708]: I1125 05:41:35.985591 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:35Z","lastTransitionTime":"2025-11-25T05:41:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.087366 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.087398 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.087452 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.087466 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.087473 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.189779 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.189819 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.189830 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.189844 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.189860 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.292174 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.292218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.292228 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.292242 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.292255 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.393654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.393768 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.393833 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.393893 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.393951 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.496142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.496171 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.496179 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.496190 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.496203 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.597980 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.598039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.598051 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.598066 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.598075 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.700209 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.700411 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.700501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.700639 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.700749 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.802557 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.802585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.802594 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.802604 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.802613 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.892661 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.892867 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:36 crc kubenswrapper[4708]: E1125 05:41:36.892878 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:36 crc kubenswrapper[4708]: E1125 05:41:36.893085 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.903797 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.903829 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.903838 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.903850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:36 crc kubenswrapper[4708]: I1125 05:41:36.903863 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:36Z","lastTransitionTime":"2025-11-25T05:41:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.005192 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.005220 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.005229 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.005241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.005250 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.106542 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.106567 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.106576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.106590 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.106599 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.207874 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.207914 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.207922 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.207936 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.207948 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.310391 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.310430 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.310438 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.310452 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.310464 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.412352 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.412387 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.412397 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.412409 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.412417 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.513889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.513914 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.513925 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.513938 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.513948 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.615675 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.615701 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.615709 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.615721 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.615729 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.717945 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.717964 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.717972 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.717982 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.717990 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.820008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.820037 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.820048 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.820059 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.820066 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.892898 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:37 crc kubenswrapper[4708]: E1125 05:41:37.893001 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.892898 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:37 crc kubenswrapper[4708]: E1125 05:41:37.893217 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.922084 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.922132 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.922145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.922159 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.922168 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:37Z","lastTransitionTime":"2025-11-25T05:41:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:37 crc kubenswrapper[4708]: I1125 05:41:37.971580 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:37 crc kubenswrapper[4708]: E1125 05:41:37.971745 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:37 crc kubenswrapper[4708]: E1125 05:41:37.971797 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:41:45.97178417 +0000 UTC m=+47.380617556 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.023591 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.023619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.023629 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.023643 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.023651 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.125167 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.125199 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.125207 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.125218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.125226 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.227316 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.227359 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.227369 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.227381 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.227389 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.328837 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.328869 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.328877 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.328888 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.328896 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.430999 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.431034 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.431044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.431056 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.431066 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.537213 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.537283 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.537297 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.537323 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.537338 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.639380 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.639411 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.639420 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.639432 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.639441 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.741034 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.741062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.741071 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.741082 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.741089 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.842510 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.842556 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.842567 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.842585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.842594 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.892837 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.892912 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:38 crc kubenswrapper[4708]: E1125 05:41:38.893031 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:38 crc kubenswrapper[4708]: E1125 05:41:38.893161 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.902679 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.911076 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.918970 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.928000 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.934720 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.943798 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.943833 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.943844 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.943857 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.943866 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:38Z","lastTransitionTime":"2025-11-25T05:41:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.946824 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.954433 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.961178 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.969271 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.977235 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.984937 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:38 crc kubenswrapper[4708]: I1125 05:41:38.994345 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.000676 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:38Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.008437 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:39Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.016771 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:39Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.028375 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:39Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.045572 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.045604 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.045613 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.045646 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.045663 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.148290 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.148332 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.148344 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.148361 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.148372 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.250815 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.250863 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.250876 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.250896 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.250913 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.353290 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.353319 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.353327 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.353339 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.353347 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.455781 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.455817 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.455827 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.455842 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.455852 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.557867 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.557941 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.557953 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.557965 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.557975 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.659778 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.659802 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.659813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.659825 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.659835 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.761397 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.761467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.761483 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.761507 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.761537 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.863131 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.863164 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.863174 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.863188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.863198 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.893006 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.893051 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:39 crc kubenswrapper[4708]: E1125 05:41:39.893126 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:39 crc kubenswrapper[4708]: E1125 05:41:39.893304 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.893994 4708 scope.go:117] "RemoveContainer" containerID="8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.965137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.965170 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.965182 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.965197 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:39 crc kubenswrapper[4708]: I1125 05:41:39.965209 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:39Z","lastTransitionTime":"2025-11-25T05:41:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.037660 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.037705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.037714 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.037730 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.037741 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.047026 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.050397 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.050422 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.050431 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.050444 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.050456 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.058612 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.061093 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.061116 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.061124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.061136 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.061144 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.068843 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.071223 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.071286 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.071299 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.071314 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.071330 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.079248 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.081836 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.081869 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.081878 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.081891 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.081901 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.089553 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.089659 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.090663 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.090687 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.090696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.090706 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.090714 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.137024 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/1.log" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.138844 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.139260 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.152024 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.163901 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.180737 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.191381 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.192627 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.192655 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.192666 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.192682 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.192693 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.208940 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.220394 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.228975 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.237696 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.248806 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.258977 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.272420 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.280812 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.289095 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.295220 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.295258 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.295272 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.295287 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.295298 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.298552 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.325654 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.338689 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:40Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.397555 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.397609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.397619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.397642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.397658 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.499913 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.499958 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.499968 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.499984 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.499999 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.602389 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.602713 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.602725 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.602738 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.602953 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.705191 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.705243 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.705255 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.705274 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.705288 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.807926 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.807960 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.807971 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.807986 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.807997 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.892900 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.893029 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.893209 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:40 crc kubenswrapper[4708]: E1125 05:41:40.893326 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.910269 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.910303 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.910311 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.910325 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:40 crc kubenswrapper[4708]: I1125 05:41:40.910335 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:40Z","lastTransitionTime":"2025-11-25T05:41:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.012113 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.012142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.012152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.012165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.012174 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.113589 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.113620 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.113628 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.113639 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.113647 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.144976 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/2.log" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.145591 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/1.log" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.147449 4708 generic.go:334] "Generic (PLEG): container finished" podID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" exitCode=1 Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.147483 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.147510 4708 scope.go:117] "RemoveContainer" containerID="8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.148014 4708 scope.go:117] "RemoveContainer" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" Nov 25 05:41:41 crc kubenswrapper[4708]: E1125 05:41:41.148144 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.158849 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.168419 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.176692 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.183721 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.192215 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.199366 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.208148 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.215407 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.215432 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.215441 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.215456 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.215467 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.217003 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.225473 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.233505 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.241003 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.247948 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.257510 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.270433 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8d6a5c60f85740c452442ce00f6769e6bb35fcc8f794242fed6b4858932f4827\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"message\\\":\\\"443 [192.168.126.11] []}]\\\\nF1125 05:41:25.746127 6140 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:25Z is after 2025-08-24T17:21:41Z]\\\\nI1125 05:41:25.746133 6140 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746131 6140 services_controller.go:444] Built service openshift-route-controller-manager/route-controller-manager LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1125 05:41:25.746141 6140 services_controller.go:444] Built service default/kubernetes LB per-node configs for network=default\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.280967 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.288875 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:41Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.317701 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.317722 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.317731 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.317744 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.317754 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.419930 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.419964 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.419974 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.419998 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.420010 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.522577 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.522613 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.522624 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.522637 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.522647 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.624342 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.624381 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.624394 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.624409 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.624419 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.726321 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.726350 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.726360 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.726372 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.726381 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.828354 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.828381 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.828390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.828401 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.828409 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.892558 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:41 crc kubenswrapper[4708]: E1125 05:41:41.892675 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.892564 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:41 crc kubenswrapper[4708]: E1125 05:41:41.892953 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.930639 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.930692 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.930701 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.930713 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:41 crc kubenswrapper[4708]: I1125 05:41:41.930722 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:41Z","lastTransitionTime":"2025-11-25T05:41:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.032797 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.032832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.032841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.032856 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.032868 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.134889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.134927 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.134936 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.134949 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.134957 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.152474 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/2.log" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.155948 4708 scope.go:117] "RemoveContainer" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" Nov 25 05:41:42 crc kubenswrapper[4708]: E1125 05:41:42.156219 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.164341 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.174833 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.183495 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.192774 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.202537 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.214433 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.221481 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.227779 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.235680 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.237266 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.237313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.237323 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.237337 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.237346 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.244449 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.252252 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.259321 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.267236 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.275300 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.282515 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.290820 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:42Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.339511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.339571 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.339580 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.339595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.339607 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.441316 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.441343 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.441354 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.441367 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.441377 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.543061 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.543099 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.543108 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.543121 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.543131 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.645261 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.645303 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.645312 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.645328 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.645338 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.747086 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.747121 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.747129 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.747141 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.747152 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.848605 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.848637 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.848645 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.848657 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.848667 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.892379 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:42 crc kubenswrapper[4708]: E1125 05:41:42.892484 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.892379 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:42 crc kubenswrapper[4708]: E1125 05:41:42.892595 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.949758 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.949792 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.949801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.949812 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:42 crc kubenswrapper[4708]: I1125 05:41:42.949821 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:42Z","lastTransitionTime":"2025-11-25T05:41:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.051750 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.051783 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.051790 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.051801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.051810 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.153994 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.154028 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.154037 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.154062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.154072 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.256392 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.256435 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.256445 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.256461 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.256470 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.358164 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.358194 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.358210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.358225 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.358234 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.459574 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.459606 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.459617 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.459630 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.459640 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.561947 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.561977 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.561986 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.562000 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.562009 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.664030 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.664074 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.664083 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.664095 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.664103 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.766220 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.766248 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.766257 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.766267 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.766274 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.868015 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.868050 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.868058 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.868068 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.868076 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.892559 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.892626 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:43 crc kubenswrapper[4708]: E1125 05:41:43.892720 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:43 crc kubenswrapper[4708]: E1125 05:41:43.892838 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.969608 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.969638 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.969651 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.969662 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:43 crc kubenswrapper[4708]: I1125 05:41:43.969669 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:43Z","lastTransitionTime":"2025-11-25T05:41:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.071240 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.071261 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.071270 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.071278 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.071285 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.172481 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.172506 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.172514 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.172539 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.172546 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.274325 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.274350 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.274358 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.274367 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.274374 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.375735 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.375767 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.375776 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.375788 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.375796 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.477649 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.477749 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.477763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.477772 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.477780 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.578810 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.578836 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.578844 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.578853 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.578860 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.682095 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.682241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.682313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.682373 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.682429 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.783975 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.783996 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.784003 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.784012 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.784019 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.885444 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.885470 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.885478 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.885488 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.885496 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.892810 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:44 crc kubenswrapper[4708]: E1125 05:41:44.892894 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.893000 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:44 crc kubenswrapper[4708]: E1125 05:41:44.893132 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.987606 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.987947 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.988014 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.988090 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:44 crc kubenswrapper[4708]: I1125 05:41:44.988153 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:44Z","lastTransitionTime":"2025-11-25T05:41:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.089773 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.089804 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.089813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.089827 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.089835 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.190940 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.190963 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.190971 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.190981 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.190988 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.292317 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.292346 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.292355 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.292364 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.292372 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.394378 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.394398 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.394406 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.394415 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.394421 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.496176 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.496200 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.496208 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.496218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.496226 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.598571 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.598600 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.598609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.598621 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.598628 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.700151 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.700188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.700198 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.700213 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.700224 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.801802 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.801835 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.801863 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.801874 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.801883 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.892318 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:45 crc kubenswrapper[4708]: E1125 05:41:45.892422 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.892580 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:45 crc kubenswrapper[4708]: E1125 05:41:45.892630 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.903584 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.903628 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.903637 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.903646 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:45 crc kubenswrapper[4708]: I1125 05:41:45.903652 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:45Z","lastTransitionTime":"2025-11-25T05:41:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.005000 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.005030 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.005038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.005049 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.005074 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.037144 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:46 crc kubenswrapper[4708]: E1125 05:41:46.037287 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:46 crc kubenswrapper[4708]: E1125 05:41:46.037332 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:02.03732038 +0000 UTC m=+63.446153766 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.106966 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.107036 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.107046 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.107057 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.107066 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.208607 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.208636 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.208647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.208656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.208664 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.310307 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.310338 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.310349 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.310362 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.310371 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.412025 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.412054 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.412062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.412072 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.412081 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.514001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.514043 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.514052 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.514062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.514069 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.615410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.615473 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.615481 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.615515 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.615545 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.716934 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.716967 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.716976 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.716990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.716999 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.819132 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.819158 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.819168 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.819179 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.819186 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.892512 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.892591 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:46 crc kubenswrapper[4708]: E1125 05:41:46.892677 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:46 crc kubenswrapper[4708]: E1125 05:41:46.892734 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.920448 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.920469 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.920478 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.920487 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:46 crc kubenswrapper[4708]: I1125 05:41:46.920496 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:46Z","lastTransitionTime":"2025-11-25T05:41:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.022739 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.022777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.022787 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.022803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.022815 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.124741 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.124777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.124805 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.124819 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.124828 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.226845 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.226875 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.226883 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.226895 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.226907 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.328813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.328841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.328850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.328862 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.328871 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.430125 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.430162 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.430174 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.430187 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.430199 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.533691 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.533723 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.533732 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.533743 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.533752 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.635807 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.635834 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.635842 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.635853 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.635862 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.737416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.737454 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.737465 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.737478 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.737487 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.839595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.839628 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.839638 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.839669 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.839680 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.892904 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.892923 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:47 crc kubenswrapper[4708]: E1125 05:41:47.892983 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:47 crc kubenswrapper[4708]: E1125 05:41:47.893140 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.941696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.941731 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.941740 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.941753 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:47 crc kubenswrapper[4708]: I1125 05:41:47.941762 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:47Z","lastTransitionTime":"2025-11-25T05:41:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.043623 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.043648 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.043657 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.043668 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.043684 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.145535 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.145565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.145574 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.145585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.145595 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.247037 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.247076 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.247085 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.247094 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.247102 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.348106 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.348128 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.348137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.348145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.348152 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.450049 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.450091 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.450100 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.450115 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.450127 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.551585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.551610 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.551619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.551630 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.551638 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.653270 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.653304 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.653315 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.653327 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.653336 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.655695 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.655843 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:20.655829475 +0000 UTC m=+82.064662861 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755749 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755785 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755795 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755808 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755819 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755946 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.755972 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.756010 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.756031 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756091 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756103 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756100 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756125 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756137 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756155 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:20.756142744 +0000 UTC m=+82.164976130 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756115 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756192 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756107 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756174 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:20.756162071 +0000 UTC m=+82.164995456 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756281 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:20.75626819 +0000 UTC m=+82.165101575 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.756298 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:20.756292154 +0000 UTC m=+82.165125540 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.857939 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.857967 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.857975 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.857994 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.858003 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.892664 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.892769 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.892670 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:48 crc kubenswrapper[4708]: E1125 05:41:48.892869 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.901424 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.909900 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.922177 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.929785 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.936919 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.945639 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.954511 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.959292 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.959319 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.959328 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.959341 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.959350 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:48Z","lastTransitionTime":"2025-11-25T05:41:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.961966 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.968582 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.976354 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.984682 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:48 crc kubenswrapper[4708]: I1125 05:41:48.995643 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:48Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.003601 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.010837 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.018468 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.024893 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.061047 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.061093 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.061102 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.061111 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.061119 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.162953 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.163006 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.163030 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.163042 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.163050 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.264491 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.264556 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.264568 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.264580 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.264589 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.366605 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.366634 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.366644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.366657 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.366666 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.468074 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.468102 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.468110 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.468120 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.468127 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.569654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.569685 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.569693 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.569723 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.569732 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.671381 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.671409 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.671416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.671425 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.671433 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.713100 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.719686 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.720265 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.729094 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.737960 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.745839 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.753943 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.761228 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.768914 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.773013 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.773035 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.773044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.773055 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.773064 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.776696 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.783075 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.794705 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.801569 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.807934 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.815888 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.823221 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.829985 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.838691 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:49Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.875047 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.875076 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.875085 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.875098 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.875107 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.892246 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.892299 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:49 crc kubenswrapper[4708]: E1125 05:41:49.892348 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:49 crc kubenswrapper[4708]: E1125 05:41:49.892465 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.976810 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.976841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.976850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.976863 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:49 crc kubenswrapper[4708]: I1125 05:41:49.976874 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:49Z","lastTransitionTime":"2025-11-25T05:41:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.078224 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.078252 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.078262 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.078273 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.078283 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.179235 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.179261 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.179272 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.179284 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.179292 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.281330 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.281386 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.281399 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.281411 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.281421 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.383295 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.383319 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.383328 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.383338 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.383347 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.471080 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.471105 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.471113 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.471124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.471132 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.479148 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:50Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.481273 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.481299 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.481308 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.481319 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.481326 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.489022 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:50Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.491398 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.491437 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.491447 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.491463 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.491472 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.498782 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:50Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.500732 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.500757 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.500765 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.500777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.500785 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.508629 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:50Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.510618 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.510647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.510656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.510665 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.510673 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.517747 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:50Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.517872 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.518780 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.518805 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.518813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.518823 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.518829 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.620483 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.620509 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.620536 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.620550 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.620560 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.722399 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.722446 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.722457 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.722467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.722476 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.824445 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.824475 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.824483 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.824494 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.824501 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.893018 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.893118 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.893126 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:50 crc kubenswrapper[4708]: E1125 05:41:50.893197 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.926212 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.926233 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.926241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.926250 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:50 crc kubenswrapper[4708]: I1125 05:41:50.926257 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:50Z","lastTransitionTime":"2025-11-25T05:41:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.028300 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.028326 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.028334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.028343 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.028351 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.130566 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.130592 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.130601 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.130612 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.130620 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.232262 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.232288 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.232296 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.232305 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.232312 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.334163 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.334190 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.334198 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.334207 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.334214 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.436396 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.436436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.436444 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.436457 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.436468 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.538552 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.538590 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.538599 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.538611 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.538619 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.640175 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.640209 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.640218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.640230 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.640239 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.742072 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.742118 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.742127 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.742145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.742154 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.843856 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.843897 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.843906 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.843922 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.843933 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.892696 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.892712 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:51 crc kubenswrapper[4708]: E1125 05:41:51.892790 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:51 crc kubenswrapper[4708]: E1125 05:41:51.892876 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.946085 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.946106 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.946114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.946124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:51 crc kubenswrapper[4708]: I1125 05:41:51.946132 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:51Z","lastTransitionTime":"2025-11-25T05:41:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.048140 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.048165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.048275 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.048289 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.048298 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.149588 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.149612 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.149619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.149631 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.149639 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.251188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.251210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.251218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.251228 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.251237 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.353020 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.353049 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.353059 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.353092 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.353100 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.454899 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.454925 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.454935 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.454956 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.454964 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.556627 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.556653 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.556661 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.556671 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.556679 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.658613 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.658649 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.658659 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.658673 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.658682 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.760627 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.760665 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.760674 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.760691 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.760701 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.862927 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.862970 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.862979 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.862991 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.863000 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.892505 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.892585 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:52 crc kubenswrapper[4708]: E1125 05:41:52.892680 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:52 crc kubenswrapper[4708]: E1125 05:41:52.892740 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.964911 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.964945 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.964954 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.964966 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:52 crc kubenswrapper[4708]: I1125 05:41:52.964974 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:52Z","lastTransitionTime":"2025-11-25T05:41:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.066943 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.066973 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.066982 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.066992 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.066999 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.168656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.168700 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.168711 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.168721 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.168729 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.270454 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.270485 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.270496 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.270512 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.270540 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.372005 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.372039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.372047 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.372058 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.372066 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.473768 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.473831 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.473841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.473852 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.473859 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.575601 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.575648 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.575664 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.575681 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.575693 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.677555 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.677579 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.677588 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.677599 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.677606 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.779472 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.779498 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.779508 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.779531 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.779540 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.881588 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.881609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.881616 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.881625 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.881634 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.892897 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.892960 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:53 crc kubenswrapper[4708]: E1125 05:41:53.892991 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:53 crc kubenswrapper[4708]: E1125 05:41:53.893063 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.983155 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.983195 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.983206 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.983219 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:53 crc kubenswrapper[4708]: I1125 05:41:53.983227 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:53Z","lastTransitionTime":"2025-11-25T05:41:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.085070 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.085128 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.085137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.085150 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.085157 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.186631 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.186656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.186663 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.186674 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.186682 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.288041 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.288073 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.288083 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.288094 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.288103 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.389400 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.389431 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.389439 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.389450 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.389459 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.491068 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.491181 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.491271 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.491356 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.491425 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.593007 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.593034 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.593042 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.593053 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.593062 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.695153 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.695183 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.695191 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.695201 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.695209 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.796483 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.796509 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.796535 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.796546 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.796553 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.892486 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:54 crc kubenswrapper[4708]: E1125 05:41:54.892627 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.892826 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:54 crc kubenswrapper[4708]: E1125 05:41:54.892896 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.893667 4708 scope.go:117] "RemoveContainer" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" Nov 25 05:41:54 crc kubenswrapper[4708]: E1125 05:41:54.893808 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.898321 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.898344 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.898355 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.898366 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:54 crc kubenswrapper[4708]: I1125 05:41:54.898375 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:54Z","lastTransitionTime":"2025-11-25T05:41:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.000462 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.000488 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.000497 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.000506 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.000514 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.102400 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.102430 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.102440 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.102451 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.102461 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.204457 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.204490 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.204501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.204513 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.204540 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.306479 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.306552 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.306562 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.306573 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.306580 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.408600 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.408658 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.408667 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.408688 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.408696 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.510279 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.510321 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.510334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.510349 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.510362 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.611990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.612026 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.612037 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.612050 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.612058 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.713390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.713418 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.713426 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.713435 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.713441 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.815424 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.815448 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.815456 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.815467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.815474 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.892558 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:55 crc kubenswrapper[4708]: E1125 05:41:55.892658 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.892569 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:55 crc kubenswrapper[4708]: E1125 05:41:55.892739 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.916565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.916591 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.916600 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.916609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:55 crc kubenswrapper[4708]: I1125 05:41:55.916617 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:55Z","lastTransitionTime":"2025-11-25T05:41:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.018380 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.018408 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.018416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.018425 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.018433 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.119705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.119739 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.119748 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.119763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.119771 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.221654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.221679 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.221687 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.221696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.221707 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.323812 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.323841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.323850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.323860 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.323867 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.425763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.425799 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.425810 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.425824 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.425836 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.527392 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.527426 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.527436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.527447 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.527456 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.628712 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.628743 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.628752 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.628764 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.628773 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.730517 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.730566 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.730576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.730586 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.730594 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.831907 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.831936 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.831946 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.831956 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.831963 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.892266 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.892290 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:56 crc kubenswrapper[4708]: E1125 05:41:56.892375 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:56 crc kubenswrapper[4708]: E1125 05:41:56.892444 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.933471 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.933494 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.933503 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.933513 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:56 crc kubenswrapper[4708]: I1125 05:41:56.933537 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:56Z","lastTransitionTime":"2025-11-25T05:41:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.035437 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.035467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.035476 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.035489 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.035499 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.137592 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.137621 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.137631 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.137642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.137651 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.239808 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.239832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.239840 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.239850 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.239859 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.341190 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.341211 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.341220 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.341230 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.341236 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.442852 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.442878 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.442898 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.442910 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.442919 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.544466 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.544500 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.544509 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.544700 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.544725 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.646972 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.647172 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.647181 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.647193 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.647203 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.748595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.748616 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.748626 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.748636 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.748645 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.850333 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.850357 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.850364 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.850374 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.850382 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.892261 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.892283 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:57 crc kubenswrapper[4708]: E1125 05:41:57.892475 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:57 crc kubenswrapper[4708]: E1125 05:41:57.892558 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.952226 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.952258 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.952266 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.952276 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:57 crc kubenswrapper[4708]: I1125 05:41:57.952285 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:57Z","lastTransitionTime":"2025-11-25T05:41:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.053979 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.054008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.054017 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.054027 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.054036 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.155448 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.155482 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.155491 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.155503 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.155513 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.256747 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.256767 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.256775 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.256783 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.256791 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.358241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.358267 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.358275 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.358285 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.358293 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.460090 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.460114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.460122 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.460133 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.460140 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.561908 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.561939 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.561949 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.561960 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.561968 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.663511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.663557 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.663565 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.663575 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.663585 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.765108 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.765139 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.765148 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.765160 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.765169 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.866438 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.866978 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.866990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.867012 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.867021 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.892231 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.892276 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:41:58 crc kubenswrapper[4708]: E1125 05:41:58.892333 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:41:58 crc kubenswrapper[4708]: E1125 05:41:58.892454 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.902545 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.910255 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.917828 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.925394 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.932153 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.941612 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.953207 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.961506 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.967846 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.968756 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.968784 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.968792 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.968803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.968812 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:58Z","lastTransitionTime":"2025-11-25T05:41:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.975801 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.984920 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.992555 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:58 crc kubenswrapper[4708]: I1125 05:41:58.999425 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:58Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.007604 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:59Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.014836 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:59Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.022252 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:59Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.030729 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:41:59Z is after 2025-08-24T17:21:41Z" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.070463 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.070493 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.070501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.070514 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.070543 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.172213 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.172251 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.172264 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.172276 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.172286 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.275851 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.275893 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.275904 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.275918 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.275927 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.378120 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.378151 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.378160 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.378171 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.378179 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.480044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.480065 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.480073 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.480083 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.480089 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.582272 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.582304 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.582312 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.582327 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.582336 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.684095 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.684126 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.684134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.684145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.684154 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.785985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.786022 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.786032 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.786044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.786055 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.888250 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.888280 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.888289 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.888303 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.888312 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.892566 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.892573 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:41:59 crc kubenswrapper[4708]: E1125 05:41:59.892687 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:41:59 crc kubenswrapper[4708]: E1125 05:41:59.892790 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.990402 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.990433 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.990458 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.990472 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:41:59 crc kubenswrapper[4708]: I1125 05:41:59.990482 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:41:59Z","lastTransitionTime":"2025-11-25T05:41:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.092285 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.092315 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.092340 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.092353 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.092361 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.194185 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.194236 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.194247 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.194258 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.194267 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.295442 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.295466 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.295474 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.295484 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.295492 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.397772 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.397824 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.397834 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.397846 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.397862 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.499751 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.499778 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.499786 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.499799 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.499809 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.601829 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.601893 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.601905 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.601915 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.601925 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.636946 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.636977 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.636985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.636996 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.637004 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.645143 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:00Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.647171 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.647201 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.647211 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.647220 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.647228 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.655129 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:00Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.656985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.657009 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.657016 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.657027 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.657035 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.664368 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:00Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.666193 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.666219 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.666228 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.666240 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.666247 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.673496 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:00Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.675394 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.675440 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.675451 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.675460 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.675467 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.682489 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:00Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.682611 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.703541 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.703570 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.703579 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.703590 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.703598 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.805650 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.805675 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.805683 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.805694 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.805703 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.892963 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.893000 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.893074 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:00 crc kubenswrapper[4708]: E1125 05:42:00.893157 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.907077 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.907098 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.907105 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.907114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:00 crc kubenswrapper[4708]: I1125 05:42:00.907122 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:00Z","lastTransitionTime":"2025-11-25T05:42:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.009106 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.009136 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.009143 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.009153 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.009160 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.110961 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.110984 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.110992 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.111001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.111008 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.213841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.213883 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.213899 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.213911 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.213920 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.315304 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.315337 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.315347 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.315357 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.315364 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.417634 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.417662 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.417670 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.417680 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.417688 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.519143 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.519175 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.519185 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.519200 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.519210 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.621134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.621158 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.621168 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.621180 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.621189 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.722979 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.723001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.723009 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.723020 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.723027 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.824803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.824843 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.824852 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.824864 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.824873 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.892291 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.892323 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:01 crc kubenswrapper[4708]: E1125 05:42:01.892377 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:01 crc kubenswrapper[4708]: E1125 05:42:01.892414 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.927036 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.927068 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.927077 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.927088 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:01 crc kubenswrapper[4708]: I1125 05:42:01.927097 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:01Z","lastTransitionTime":"2025-11-25T05:42:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.029180 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.029214 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.029224 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.029237 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.029246 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.065473 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:02 crc kubenswrapper[4708]: E1125 05:42:02.065574 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:42:02 crc kubenswrapper[4708]: E1125 05:42:02.065625 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:34.06561179 +0000 UTC m=+95.474445176 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.131328 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.131389 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.131397 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.131410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.131419 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.232652 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.232686 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.232697 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.232712 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.232722 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.334705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.334726 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.334735 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.334744 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.334752 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.436485 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.436513 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.436539 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.436549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.436556 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.538462 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.538488 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.538496 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.538507 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.538529 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.640677 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.640705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.640713 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.640723 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.640730 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.742512 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.742637 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.742695 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.742768 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.742839 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.844785 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.844848 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.844860 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.844870 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.844877 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.892663 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:02 crc kubenswrapper[4708]: E1125 05:42:02.892853 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.892869 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:02 crc kubenswrapper[4708]: E1125 05:42:02.893082 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.946577 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.946610 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.946619 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.946629 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:02 crc kubenswrapper[4708]: I1125 05:42:02.946638 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:02Z","lastTransitionTime":"2025-11-25T05:42:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.048680 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.048707 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.048715 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.048742 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.048751 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.150643 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.150671 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.150680 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.150689 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.150699 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.252601 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.252625 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.252635 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.252647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.252654 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.354770 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.354795 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.354803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.354815 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.354832 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.456464 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.456486 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.456495 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.456505 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.456512 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.558740 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.558768 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.558778 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.558790 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.558800 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.660103 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.660137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.660145 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.660157 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.660172 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.761307 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.761337 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.761346 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.761359 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.761367 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.863543 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.863575 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.863583 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.863595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.863606 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.892199 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.892248 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:03 crc kubenswrapper[4708]: E1125 05:42:03.892309 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:03 crc kubenswrapper[4708]: E1125 05:42:03.892451 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.965237 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.965267 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.965276 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.965303 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:03 crc kubenswrapper[4708]: I1125 05:42:03.965312 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:03Z","lastTransitionTime":"2025-11-25T05:42:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.067325 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.067359 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.067368 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.067380 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.067389 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.168867 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.168891 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.168899 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.168909 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.168916 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.206977 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rjrnw_abcfd1aa-ec01-4cae-8858-0d7b140c6344/kube-multus/0.log" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.207016 4708 generic.go:334] "Generic (PLEG): container finished" podID="abcfd1aa-ec01-4cae-8858-0d7b140c6344" containerID="cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1" exitCode=1 Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.207056 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerDied","Data":"cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.207445 4708 scope.go:117] "RemoveContainer" containerID="cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.214787 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.230332 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.238072 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.244981 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.253030 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.261217 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.269409 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.271108 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.271131 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.271140 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.271152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.271160 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.279664 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.288956 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.295604 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.303190 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.311446 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.318803 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.326389 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.334210 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.341831 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.349645 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:04Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:04Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.372858 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.372884 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.372892 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.372905 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.372913 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.474252 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.474287 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.474297 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.474309 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.474317 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.575846 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.575871 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.575881 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.575891 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.575899 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.677910 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.677976 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.677990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.678003 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.678012 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.779845 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.779871 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.779880 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.779891 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.779901 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.881070 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.881106 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.881115 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.881128 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.881138 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.892830 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.892853 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:04 crc kubenswrapper[4708]: E1125 05:42:04.892919 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:04 crc kubenswrapper[4708]: E1125 05:42:04.892983 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.983127 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.983151 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.983160 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.983171 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:04 crc kubenswrapper[4708]: I1125 05:42:04.983179 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:04Z","lastTransitionTime":"2025-11-25T05:42:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.085014 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.085039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.085048 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.085059 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.085067 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.186005 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.186030 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.186038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.186050 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.186057 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.209927 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rjrnw_abcfd1aa-ec01-4cae-8858-0d7b140c6344/kube-multus/0.log" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.209980 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerStarted","Data":"78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.219217 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.226338 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.234493 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.243810 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.251818 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.259934 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.266710 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.274363 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.280705 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.287655 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.288106 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.288137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.288149 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.288161 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.288169 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.296514 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.307959 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.315011 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.321536 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.332711 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.341215 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.348950 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:05Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.390135 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.390157 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.390165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.390176 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.390184 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.491624 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.491652 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.491662 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.491674 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.491683 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.593289 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.593311 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.593319 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.593331 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.593339 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.694512 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.694636 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.694704 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.694763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.694845 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.795985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.796017 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.796026 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.796039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.796048 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.892663 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.892735 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:05 crc kubenswrapper[4708]: E1125 05:42:05.892988 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:05 crc kubenswrapper[4708]: E1125 05:42:05.893064 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.893126 4708 scope.go:117] "RemoveContainer" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.897317 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.897343 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.897353 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.897364 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.897373 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.998409 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.998433 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.998441 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.998453 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:05 crc kubenswrapper[4708]: I1125 05:42:05.998462 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:05Z","lastTransitionTime":"2025-11-25T05:42:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.100399 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.100428 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.100439 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.100450 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.100459 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.201994 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.202024 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.202033 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.202046 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.202054 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.213660 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/2.log" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.216412 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.216756 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.229085 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.239620 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.247782 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.254408 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.263173 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.270767 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.278348 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.286709 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.293767 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.301536 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.303499 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.303536 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.303545 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.303558 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.303567 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.308808 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.316149 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.324364 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.332096 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.341190 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.364086 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.381243 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:06Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.405491 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.405536 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.405545 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.405558 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.405566 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.507738 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.507772 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.507781 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.507804 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.507813 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.609459 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.609494 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.609505 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.609535 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.609545 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.710986 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.711008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.711016 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.711027 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.711037 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.812909 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.813044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.813118 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.813209 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.813281 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.893231 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.893259 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:06 crc kubenswrapper[4708]: E1125 05:42:06.893473 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:06 crc kubenswrapper[4708]: E1125 05:42:06.893391 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.914610 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.914647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.914657 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.914669 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:06 crc kubenswrapper[4708]: I1125 05:42:06.914678 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:06Z","lastTransitionTime":"2025-11-25T05:42:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.015898 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.015925 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.015934 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.015945 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.015954 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.117664 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.117695 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.117703 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.117714 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.117722 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.218907 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.218940 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.218949 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.218958 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.218965 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.221050 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/3.log" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.221592 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/2.log" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.223809 4708 generic.go:334] "Generic (PLEG): container finished" podID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" exitCode=1 Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.223840 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.223985 4708 scope.go:117] "RemoveContainer" containerID="bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.224254 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:42:07 crc kubenswrapper[4708]: E1125 05:42:07.224462 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.233210 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.242954 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.252810 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.261465 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.270897 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.279312 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.287580 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.295970 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.303868 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.320340 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.320371 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.320380 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.320391 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.320399 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.325717 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bc769d9f18c90d0c66c0824208829ab5aa6ac6b5a451a9e48ae54040840378de\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:41:40Z\\\",\\\"message\\\":\\\"ultus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.580244 6378 default_network_controller.go:776] Recording success event on pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 05:41:40.580869 6378 services_controller.go:356] Processing sync for service openshift-ingress-operator/metrics for network=default\\\\nI1125 05:41:40.581387 6378 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1125 05:41:40.581412 6378 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-additional-cni-plugins-rx7k7 in node crc\\\\nI1125 05:41:40.581423 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-additional-cni-plugins-rx7k7 after 0 failed attempt(s)\\\\nI1125 05:41:40.581428 6378 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-rx7k7\\\\nI1125 05:41:40.581416 6378 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-node-identity/network-node-identity-vrzqb after 0 failed attempt(s)\\\\nI1125 05:41:40.581468 6378 default_network_controller.go:776] Recording success event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nF1125 05:41:40.581489 6378 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:06Z\\\",\\\"message\\\":\\\"9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 05:42:06.498470 6752 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"7594bb65-e742-44b3-a975-d639b1128be5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:42:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.333385 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.340172 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.350157 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.359299 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.367701 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.375725 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.386446 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:07Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.422264 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.422311 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.422321 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.422344 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.422358 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.524484 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.524605 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.524673 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.524741 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.524835 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.626415 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.626556 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.626628 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.626704 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.626768 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.728817 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.728861 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.728876 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.728897 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.728910 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.830614 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.830644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.830656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.830671 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.830683 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.892639 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.892679 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:07 crc kubenswrapper[4708]: E1125 05:42:07.892758 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:07 crc kubenswrapper[4708]: E1125 05:42:07.892920 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.932069 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.932102 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.932114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.932132 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:07 crc kubenswrapper[4708]: I1125 05:42:07.932144 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:07Z","lastTransitionTime":"2025-11-25T05:42:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.034349 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.034388 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.034399 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.034415 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.034425 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.136362 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.136400 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.136410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.136423 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.136434 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.229030 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/3.log" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.231757 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:42:08 crc kubenswrapper[4708]: E1125 05:42:08.231889 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.238407 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.238437 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.238448 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.238461 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.238471 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.244408 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.255211 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.265728 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.277139 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.291464 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:06Z\\\",\\\"message\\\":\\\"9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 05:42:06.498470 6752 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"7594bb65-e742-44b3-a975-d639b1128be5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:42:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.301878 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.310206 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.319602 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.330704 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.340542 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.340572 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.340584 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.340599 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.340611 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.341486 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.349959 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.357257 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.365661 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.375213 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.384269 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.393260 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.400403 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.442159 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.442191 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.442200 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.442213 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.442223 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.544478 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.544610 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.544696 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.544784 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.544854 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.649172 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.649352 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.649427 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.649511 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.649617 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.751599 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.751636 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.751647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.751661 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.751670 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.853738 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.853777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.853787 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.853809 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.853822 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.892469 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.892625 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:08 crc kubenswrapper[4708]: E1125 05:42:08.892801 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:08 crc kubenswrapper[4708]: E1125 05:42:08.892943 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.903118 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.917492 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:06Z\\\",\\\"message\\\":\\\"9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 05:42:06.498470 6752 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"7594bb65-e742-44b3-a975-d639b1128be5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:42:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.925209 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.933165 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.943531 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.955656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.955689 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.955698 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.955714 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.955725 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:08Z","lastTransitionTime":"2025-11-25T05:42:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.957800 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.969028 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.978904 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:08 crc kubenswrapper[4708]: I1125 05:42:08.991012 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:08Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.004670 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.013686 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.024190 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.032695 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.041490 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.050443 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.058160 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.058204 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.058218 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.058239 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.058254 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.059301 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.068633 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:09Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.160326 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.160378 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.160389 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.160403 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.160415 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.262402 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.262424 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.262435 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.262447 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.262455 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.364311 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.364344 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.364354 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.364362 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.364371 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.466654 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.466693 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.466705 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.466721 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.466736 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.568932 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.568963 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.568976 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.568993 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.569005 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.670775 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.670820 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.670832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.670856 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.670870 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.774734 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.774801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.774813 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.774829 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.774843 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.877044 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.877087 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.877099 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.877115 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.877127 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.892735 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.892789 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:09 crc kubenswrapper[4708]: E1125 05:42:09.892876 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:09 crc kubenswrapper[4708]: E1125 05:42:09.892966 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.978617 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.978718 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.978889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.979039 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:09 crc kubenswrapper[4708]: I1125 05:42:09.979178 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:09Z","lastTransitionTime":"2025-11-25T05:42:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.080842 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.081000 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.081076 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.081154 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.081210 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.183028 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.183062 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.183072 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.183086 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.183096 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.284730 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.284828 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.284908 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.284978 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.285049 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.387503 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.387563 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.387595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.387611 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.387623 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.489780 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.489814 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.489823 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.489834 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.489843 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.591550 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.591574 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.591582 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.591595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.591604 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.693936 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.693978 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.693988 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.694008 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.694023 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.796478 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.796508 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.796543 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.796555 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.796564 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.809419 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.809442 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.809450 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.809462 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.809469 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.818393 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:10Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.824932 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.824963 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.824974 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.824985 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.824993 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.836503 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:10Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.839102 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.839137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.839147 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.839158 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.839166 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.847698 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:10Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.849988 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.850070 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.850155 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.850217 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.850278 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.858442 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:10Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.860721 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.860828 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.860890 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.860942 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.860994 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.868700 4708 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2c611337-7e15-4e6b-a515-a41d6885899c\\\",\\\"systemUUID\\\":\\\"62ba027d-f75d-44c0-93ed-f3312d588154\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:10Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.868939 4708 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.893050 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.893121 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.893172 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:10 crc kubenswrapper[4708]: E1125 05:42:10.893205 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.898248 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.898334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.898410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.898477 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:10 crc kubenswrapper[4708]: I1125 05:42:10.898563 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:10Z","lastTransitionTime":"2025-11-25T05:42:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.000090 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.000123 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.000134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.000148 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.000157 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.101963 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.101990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.102001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.102015 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.102023 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.203990 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.204107 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.204174 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.204244 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.204305 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.305647 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.305684 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.305695 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.305710 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.305720 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.407783 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.407807 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.407925 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.407941 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.407951 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.509436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.509465 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.509476 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.509493 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.509504 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.611449 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.611480 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.611489 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.611500 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.611507 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.713686 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.713710 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.713720 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.713733 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.713742 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.815504 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.815642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.815707 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.815784 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.815852 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.893041 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.893052 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:11 crc kubenswrapper[4708]: E1125 05:42:11.893311 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:11 crc kubenswrapper[4708]: E1125 05:42:11.893213 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.900270 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.917422 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.917546 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.917615 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.917675 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:11 crc kubenswrapper[4708]: I1125 05:42:11.917731 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:11Z","lastTransitionTime":"2025-11-25T05:42:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.019242 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.019433 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.019493 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.019593 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.019663 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.121411 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.121568 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.121629 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.121703 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.121785 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.223471 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.223505 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.223514 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.223552 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.223563 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.325751 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.325791 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.325800 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.325814 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.325824 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.427309 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.427437 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.427501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.427597 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.427668 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.529286 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.529316 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.529325 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.529337 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.529347 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.630508 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.630671 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.630735 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.630812 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.630879 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.735709 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.737390 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.737404 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.737423 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.737434 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.839701 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.839747 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.839771 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.839786 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.839796 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.892687 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.892740 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:12 crc kubenswrapper[4708]: E1125 05:42:12.892797 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:12 crc kubenswrapper[4708]: E1125 05:42:12.892861 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.941831 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.941868 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.941878 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.941889 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:12 crc kubenswrapper[4708]: I1125 05:42:12.941898 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:12Z","lastTransitionTime":"2025-11-25T05:42:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.043703 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.043736 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.043745 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.043755 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.043773 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.145387 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.145424 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.145436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.145446 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.145454 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.247210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.247237 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.247246 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.247286 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.247296 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.349767 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.349803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.349812 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.349827 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.349835 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.451436 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.451472 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.451481 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.451494 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.451502 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.553656 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.553694 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.553725 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.553739 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.553747 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.655749 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.655794 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.655803 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.655816 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.655824 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.758204 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.758236 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.758246 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.758258 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.758266 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.859713 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.859746 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.859754 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.859777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.859787 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.892714 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.892787 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:13 crc kubenswrapper[4708]: E1125 05:42:13.892817 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:13 crc kubenswrapper[4708]: E1125 05:42:13.892898 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.960701 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.960726 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.960734 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.960743 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:13 crc kubenswrapper[4708]: I1125 05:42:13.960751 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:13Z","lastTransitionTime":"2025-11-25T05:42:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.062552 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.062580 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.062604 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.062614 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.062622 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.164278 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.164317 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.164325 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.164335 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.164342 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.266174 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.266201 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.266208 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.266217 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.266223 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.368098 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.368134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.368142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.368155 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.368164 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.469688 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.469771 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.469780 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.469795 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.469803 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.571818 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.571840 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.571849 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.571859 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.571867 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.673267 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.673296 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.673304 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.673314 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.673323 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.775164 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.775188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.775196 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.775205 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.775212 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.876548 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.876571 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.876578 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.876587 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.876594 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.893017 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.893058 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:14 crc kubenswrapper[4708]: E1125 05:42:14.893111 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:14 crc kubenswrapper[4708]: E1125 05:42:14.893156 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.978392 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.978418 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.978429 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.978440 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:14 crc kubenswrapper[4708]: I1125 05:42:14.978447 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:14Z","lastTransitionTime":"2025-11-25T05:42:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.080379 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.080405 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.080413 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.080421 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.080428 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.182368 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.182393 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.182407 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.182416 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.182423 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.284210 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.284237 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.284244 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.284254 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.284263 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.386188 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.386232 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.386241 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.386250 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.386258 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.488458 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.488501 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.488514 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.488549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.488561 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.590031 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.590054 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.590063 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.590072 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.590080 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.692354 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.692386 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.692396 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.692410 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.692419 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.794726 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.794794 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.794806 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.794832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.794851 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.892928 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.892996 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:15 crc kubenswrapper[4708]: E1125 05:42:15.893091 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:15 crc kubenswrapper[4708]: E1125 05:42:15.893273 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.897113 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.897143 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.897152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.897165 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.897174 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.998821 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.998843 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.998852 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.998862 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:15 crc kubenswrapper[4708]: I1125 05:42:15.998872 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:15Z","lastTransitionTime":"2025-11-25T05:42:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.100089 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.100119 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.100128 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.100141 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.100151 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.201878 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.201908 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.201915 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.201926 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.201936 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.303857 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.303886 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.303895 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.303907 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.303916 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.405204 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.405234 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.405251 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.405262 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.405270 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.507334 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.507365 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.507374 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.507385 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.507392 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.610085 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.610124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.610134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.610152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.610164 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.712238 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.712280 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.712293 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.712308 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.712318 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.814440 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.814473 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.814482 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.814495 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.814503 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.893000 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.893039 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:16 crc kubenswrapper[4708]: E1125 05:42:16.893109 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:16 crc kubenswrapper[4708]: E1125 05:42:16.893205 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.916086 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.916114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.916142 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.916154 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:16 crc kubenswrapper[4708]: I1125 05:42:16.916162 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:16Z","lastTransitionTime":"2025-11-25T05:42:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.017980 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.018012 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.018022 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.018033 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.018049 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.119801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.119824 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.119831 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.119841 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.119848 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.223001 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.223038 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.223046 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.223059 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.223068 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.324779 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.324816 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.324826 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.324844 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.324853 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.426269 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.426313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.426322 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.426338 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.426350 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.528932 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.528967 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.528977 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.528991 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.529000 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.630231 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.630265 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.630274 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.630287 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.630296 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.731642 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.731678 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.731687 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.731702 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.731711 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.833571 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.833603 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.833613 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.833626 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.833634 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.892482 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:17 crc kubenswrapper[4708]: E1125 05:42:17.892594 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.892697 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:17 crc kubenswrapper[4708]: E1125 05:42:17.892757 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.936148 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.936193 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.936204 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.936219 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:17 crc kubenswrapper[4708]: I1125 05:42:17.936230 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:17Z","lastTransitionTime":"2025-11-25T05:42:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.038415 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.038456 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.038467 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.038484 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.038495 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.140285 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.140321 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.140329 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.140342 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.140350 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.242781 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.242840 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.242853 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.242871 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.242882 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.345255 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.345291 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.345301 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.345313 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.345321 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.447545 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.447576 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.447586 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.447597 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.447604 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.549595 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.549622 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.549632 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.549643 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.549651 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.651589 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.651644 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.651657 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.651679 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.651695 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.753921 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.753973 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.753983 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.753993 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.754002 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.855861 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.855901 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.855912 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.855925 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.855939 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.892966 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:18 crc kubenswrapper[4708]: E1125 05:42:18.893095 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.893153 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:18 crc kubenswrapper[4708]: E1125 05:42:18.893290 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.901827 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-5wgns" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b5811ead-bd59-4da5-a686-bc1c974f86eb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c762b405204273f7f5e14e8e203e733cd73844bb66391798101c9ff00c5d92d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9fqm8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:19Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-5wgns\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.910473 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.918332 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f24da5a0-0c94-4513-9c44-84818f240046\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7678e64ff89c477ebb947174d6d705b79735cfd3e61646863613415a436196e0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h25bh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-lh62q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.928355 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e4dfd5bc-2e9a-4b1c-a99b-526d5c58be1d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7d893b130ec6e03f56ec55760b8b27edc2981218e259c30d37f37d3e8d210ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://515edd23d40efce476aecfff8f547e4352528c49174d3e08fedb74c5c1f739fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb080249b7716926a523f71209fd9a57949389c8e2b0b5317189425ba2775846\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d67d44c0f6b628f435610b9dcb162e8df8eda4a5b94d6fed9fd9d5b50e99a252\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://92c796693ad454ce4c40cda8526787a8156c2435c1b21767bacc71e534131c1d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa228325fa4b5d55b5f040db486d4bcf63ec4ad263c7bd0c54fd6d0e4dd9ca58\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:22Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c1cd77c54be29bc97920c6097cd4d57429b6717803573edff848876079f0844c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mw4xn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-rx7k7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.940261 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d78be28-7e93-4fda-98cd-521398f9b3e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:06Z\\\",\\\"message\\\":\\\"9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 05:42:06.498470 6752 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-network-diagnostics/network-check-target_TCP_cluster\\\\\\\", UUID:\\\\\\\"7594bb65-e742-44b3-a975-d639b1128be5\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-network-diagnostics/network-check-target\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"cluster\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:42:05Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7wnr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qppcd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.948316 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1da4d69-b614-43c1-ab20-37ac4b7e1e11\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f257e448818f2e1bb77d6004e0a2e4a8e29e292b1666f7a4d66e0ff1a8df40b1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1dad5e41656bc98d1dd43f6ec0c10f26f1b2c7e913df22e9dfb68c5aa6c3446\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-swwd7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:29Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-nq7ht\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.955887 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-g6lks" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"711da2a4-b5f2-40ce-84d4-619c4e3f2522\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:30Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h9fxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:30Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-g6lks\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.957492 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.957597 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.957665 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.957728 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.957803 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:18Z","lastTransitionTime":"2025-11-25T05:42:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.969404 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"293e9a7b-310c-45ea-967a-aecc41721d09\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a5ff8ba730aab123035f7bae9fe85796d4363f8fca5fa19f3427c30137974a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3557e7a583f3a88283e9fd2f09423fd27727314f49e5ded8a2fc42a572bb6290\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://780cb0f5f63f2ad2082e4f81ba79dfc234236a94a9803ab695a0a8a44a6d9fdc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://38a5204d4f5fabfd1eaa7d1877ba0cc84ae83dae4de35e2c0d1b39e50b50d2ec\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.977812 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fc9ca524dbb9f3b20e021d690bd2452cc80a8351f477733df7a58d9a96113bdd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.986989 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\" 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1125 05:41:16.271222 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI1125 05:41:16.271238 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nI1125 05:41:16.271310 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1125 05:41:16.271315 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1125 05:41:16.271335 1 tlsconfig.go:203] \\\\\\\"Loaded serving cert\\\\\\\" certName=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"localhost\\\\\\\\\\\\\\\" [serving] validServingFor=[localhost] issuer=\\\\\\\\\\\\\\\"check-endpoints-signer@1764049260\\\\\\\\\\\\\\\" (2025-11-25 05:40:59 +0000 UTC to 2025-12-25 05:41:00 +0000 UTC (now=2025-11-25 05:41:16.271316575 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271476 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1764049271\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1764049270\\\\\\\\\\\\\\\" (2025-11-25 04:41:10 +0000 UTC to 2026-11-25 04:41:10 +0000 UTC (now=2025-11-25 05:41:16.27146354 +0000 UTC))\\\\\\\"\\\\nI1125 05:41:16.271496 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI1125 05:41:16.271535 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI1125 05:41:16.271553 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-730857307/tls.crt::/tmp/serving-cert-730857307/tls.key\\\\\\\"\\\\nI1125 05:41:16.271699 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nF1125 05:41:16.271791 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:18 crc kubenswrapper[4708]: I1125 05:42:18.993643 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8d4ba54-b7e7-4bec-9863-7f8c394129d0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c91c363f72c1eec026fb1f559609070ac7660e311f094db141b407d7fbff2cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://874a6d7983c65f13276ab2f86fac9ecdfd4753473ca55f2c6b4cc4887bf17ba1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://874a6d7983c65f13276ab2f86fac9ecdfd4753473ca55f2c6b4cc4887bf17ba1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:40:59Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:18Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.001738 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.008715 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-clxd5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"917c8c88-a0de-49af-9388-990df3c83b25\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95232be612d645ebf4916944ee9bdefbcf2732c5647339e5be54b11ab6800757\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hcc7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:16Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-clxd5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.016808 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfd21e5a-2f95-4e1d-8227-afd1ff0f89f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:40:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dd62076a558295d182c8200cc95a5e05c678266306bf86de622d6792d1565071\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:40:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f5b5254b137d06f603294901270e3e3e76fe9694efab86585e11c0e92467e3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e90e00074910ddfb51091939a33614de1970ae82d22404ea335dd542449cc85f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:40:58Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.024603 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:16Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.032657 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-rjrnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"abcfd1aa-ec01-4cae-8858-0d7b140c6344\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T05:42:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T05:42:03Z\\\",\\\"message\\\":\\\"2025-11-25T05:41:18+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9\\\\n2025-11-25T05:41:18+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_863be93c-c1a3-4525-a470-2feb6c77baa9 to /host/opt/cni/bin/\\\\n2025-11-25T05:41:18Z [verbose] multus-daemon started\\\\n2025-11-25T05:41:18Z [verbose] Readiness Indicator file check\\\\n2025-11-25T05:42:03Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:42:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-w5fqd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T05:41:17Z\\\"}}\" for pod \"openshift-multus\"/\"multus-rjrnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.040189 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0acc9cb98cc8cb3477ec4a7569fa2a3bc33d1f50e54216c8f1effe4dd6ed358a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c7b39bf428011a7386b418e5fb200de798d55d1802f1bc1eafad0c436524ecdb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.047448 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:41:20Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0d12191648812f9830733b793e1484a8c1b248d4299e025b7831e954d4a6ed14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T05:41:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T05:42:19Z is after 2025-08-24T17:21:41Z" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.060180 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.060212 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.060224 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.060242 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.060252 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.161818 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.161851 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.161866 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.161879 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.161890 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.264061 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.264093 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.264101 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.264112 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.264120 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.365719 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.365767 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.365777 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.365793 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.365806 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.467769 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.467801 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.467811 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.467832 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.467845 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.569717 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.569752 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.569763 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.569773 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.569781 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.671556 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.671585 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.671597 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.671609 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.671618 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.773307 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.773331 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.773340 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.773351 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.773359 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.874438 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.874484 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.874497 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.874541 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.874553 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.892772 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:19 crc kubenswrapper[4708]: E1125 05:42:19.893027 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.893353 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:19 crc kubenswrapper[4708]: E1125 05:42:19.893557 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.976315 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.976355 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.976367 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.976392 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:19 crc kubenswrapper[4708]: I1125 05:42:19.976406 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:19Z","lastTransitionTime":"2025-11-25T05:42:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.079134 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.079177 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.079189 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.079207 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.079219 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.181332 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.181374 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.181402 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.181417 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.181427 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.283896 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.283933 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.283942 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.283954 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.283963 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.386045 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.386077 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.386089 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.386113 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.386123 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.488136 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.488158 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.488168 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.488177 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.488187 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.590080 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.590114 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.590124 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.590137 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.590148 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.692087 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.692150 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.692161 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.692177 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.692188 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.716673 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.716866 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:43:24.716841933 +0000 UTC m=+146.125675329 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.794105 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.794140 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.794152 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.794166 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.794177 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.817813 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.817857 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.817885 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.817903 4708 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.817966 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:43:24.817948438 +0000 UTC m=+146.226781824 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.817910 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.817998 4708 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818050 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 05:43:24.818039259 +0000 UTC m=+146.226872644 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818068 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818070 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818085 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818097 4708 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818109 4708 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818116 4708 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818162 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 05:43:24.818145789 +0000 UTC m=+146.226979175 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.818181 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 05:43:24.818174042 +0000 UTC m=+146.227007428 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.876247 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.876308 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.876322 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.876343 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.876360 4708 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T05:42:20Z","lastTransitionTime":"2025-11-25T05:42:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.892883 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.892984 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.892883 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.893147 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.912934 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc"] Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.913275 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:20 crc kubenswrapper[4708]: W1125 05:42:20.914551 4708 reflector.go:561] object-"openshift-cluster-version"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.914592 4708 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 05:42:20 crc kubenswrapper[4708]: W1125 05:42:20.914603 4708 reflector.go:561] object-"openshift-cluster-version"/"default-dockercfg-gxtc4": failed to list *v1.Secret: secrets "default-dockercfg-gxtc4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.914629 4708 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"default-dockercfg-gxtc4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-gxtc4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 05:42:20 crc kubenswrapper[4708]: W1125 05:42:20.914561 4708 reflector.go:561] object-"openshift-cluster-version"/"cluster-version-operator-serving-cert": failed to list *v1.Secret: secrets "cluster-version-operator-serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.914650 4708 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-version-operator-serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 05:42:20 crc kubenswrapper[4708]: W1125 05:42:20.914660 4708 reflector.go:561] object-"openshift-cluster-version"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Nov 25 05:42:20 crc kubenswrapper[4708]: E1125 05:42:20.914681 4708 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.965678 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-rjrnw" podStartSLOduration=64.965667561 podStartE2EDuration="1m4.965667561s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:20.955146649 +0000 UTC m=+82.363980035" watchObservedRunningTime="2025-11-25 05:42:20.965667561 +0000 UTC m=+82.374500947" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.971981 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-5wgns" podStartSLOduration=64.971969166 podStartE2EDuration="1m4.971969166s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:20.971885628 +0000 UTC m=+82.380719014" watchObservedRunningTime="2025-11-25 05:42:20.971969166 +0000 UTC m=+82.380802552" Nov 25 05:42:20 crc kubenswrapper[4708]: I1125 05:42:20.997505 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podStartSLOduration=64.997484526 podStartE2EDuration="1m4.997484526s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:20.996920054 +0000 UTC m=+82.405753440" watchObservedRunningTime="2025-11-25 05:42:20.997484526 +0000 UTC m=+82.406317913" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.010079 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-rx7k7" podStartSLOduration=65.010056056 podStartE2EDuration="1m5.010056056s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.009786759 +0000 UTC m=+82.418620145" watchObservedRunningTime="2025-11-25 05:42:21.010056056 +0000 UTC m=+82.418889442" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.019954 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.019993 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a715d87-7d8c-4383-a2db-bad469ae7528-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.020050 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.020172 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a715d87-7d8c-4383-a2db-bad469ae7528-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.020222 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.052260 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-nq7ht" podStartSLOduration=64.052239753 podStartE2EDuration="1m4.052239753s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.051720827 +0000 UTC m=+82.460554212" watchObservedRunningTime="2025-11-25 05:42:21.052239753 +0000 UTC m=+82.461073139" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.086536 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=32.08649779 podStartE2EDuration="32.08649779s" podCreationTimestamp="2025-11-25 05:41:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.075007093 +0000 UTC m=+82.483840479" watchObservedRunningTime="2025-11-25 05:42:21.08649779 +0000 UTC m=+82.495331176" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.097070 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=65.097048839 podStartE2EDuration="1m5.097048839s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.096479318 +0000 UTC m=+82.505312704" watchObservedRunningTime="2025-11-25 05:42:21.097048839 +0000 UTC m=+82.505882226" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.107865 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=65.107844758 podStartE2EDuration="1m5.107844758s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.107681481 +0000 UTC m=+82.516514868" watchObservedRunningTime="2025-11-25 05:42:21.107844758 +0000 UTC m=+82.516678145" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.114606 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=10.114592011 podStartE2EDuration="10.114592011s" podCreationTimestamp="2025-11-25 05:42:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.114021629 +0000 UTC m=+82.522855025" watchObservedRunningTime="2025-11-25 05:42:21.114592011 +0000 UTC m=+82.523425397" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121432 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121480 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121501 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a715d87-7d8c-4383-a2db-bad469ae7528-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121574 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a715d87-7d8c-4383-a2db-bad469ae7528-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121614 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.121725 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5a715d87-7d8c-4383-a2db-bad469ae7528-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.129892 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-clxd5" podStartSLOduration=65.129880212 podStartE2EDuration="1m5.129880212s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:21.129853241 +0000 UTC m=+82.538686628" watchObservedRunningTime="2025-11-25 05:42:21.129880212 +0000 UTC m=+82.538713598" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.816932 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.827075 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5a715d87-7d8c-4383-a2db-bad469ae7528-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.889774 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.893281 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.893281 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:21 crc kubenswrapper[4708]: E1125 05:42:21.893660 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.893881 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:42:21 crc kubenswrapper[4708]: E1125 05:42:21.893889 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:21 crc kubenswrapper[4708]: E1125 05:42:21.894038 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:42:21 crc kubenswrapper[4708]: I1125 05:42:21.895985 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a715d87-7d8c-4383-a2db-bad469ae7528-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:22 crc kubenswrapper[4708]: E1125 05:42:22.122743 4708 configmap.go:193] Couldn't get configMap openshift-cluster-version/openshift-service-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 25 05:42:22 crc kubenswrapper[4708]: E1125 05:42:22.122966 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca podName:5a715d87-7d8c-4383-a2db-bad469ae7528 nodeName:}" failed. No retries permitted until 2025-11-25 05:42:22.622944893 +0000 UTC m=+84.031778280 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca" (UniqueName: "kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca") pod "cluster-version-operator-5c965bbfc6-nngxc" (UID: "5a715d87-7d8c-4383-a2db-bad469ae7528") : failed to sync configmap cache: timed out waiting for the condition Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.159127 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.446070 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.633249 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.634223 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5a715d87-7d8c-4383-a2db-bad469ae7528-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nngxc\" (UID: \"5a715d87-7d8c-4383-a2db-bad469ae7528\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.724745 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" Nov 25 05:42:22 crc kubenswrapper[4708]: W1125 05:42:22.736605 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5a715d87_7d8c_4383_a2db_bad469ae7528.slice/crio-229bfa3dd0ee9027bb23831f52d6c8dcf16f9ef01f7cf4d7456e983d6d5934cf WatchSource:0}: Error finding container 229bfa3dd0ee9027bb23831f52d6c8dcf16f9ef01f7cf4d7456e983d6d5934cf: Status 404 returned error can't find the container with id 229bfa3dd0ee9027bb23831f52d6c8dcf16f9ef01f7cf4d7456e983d6d5934cf Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.893116 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:22 crc kubenswrapper[4708]: E1125 05:42:22.893674 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:22 crc kubenswrapper[4708]: I1125 05:42:22.893210 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:22 crc kubenswrapper[4708]: E1125 05:42:22.893934 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:23 crc kubenswrapper[4708]: I1125 05:42:23.266873 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" event={"ID":"5a715d87-7d8c-4383-a2db-bad469ae7528","Type":"ContainerStarted","Data":"512234ddad5a5073187f8c546c75640267c4dfe4b389a40de8e0d4974b2fc589"} Nov 25 05:42:23 crc kubenswrapper[4708]: I1125 05:42:23.266926 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" event={"ID":"5a715d87-7d8c-4383-a2db-bad469ae7528","Type":"ContainerStarted","Data":"229bfa3dd0ee9027bb23831f52d6c8dcf16f9ef01f7cf4d7456e983d6d5934cf"} Nov 25 05:42:23 crc kubenswrapper[4708]: I1125 05:42:23.279046 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nngxc" podStartSLOduration=67.27903177499999 podStartE2EDuration="1m7.279031775s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:23.278686316 +0000 UTC m=+84.687519702" watchObservedRunningTime="2025-11-25 05:42:23.279031775 +0000 UTC m=+84.687865161" Nov 25 05:42:23 crc kubenswrapper[4708]: I1125 05:42:23.892686 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:23 crc kubenswrapper[4708]: E1125 05:42:23.893104 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:23 crc kubenswrapper[4708]: I1125 05:42:23.892692 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:23 crc kubenswrapper[4708]: E1125 05:42:23.893303 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:24 crc kubenswrapper[4708]: I1125 05:42:24.893471 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:24 crc kubenswrapper[4708]: I1125 05:42:24.893469 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:24 crc kubenswrapper[4708]: E1125 05:42:24.893641 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:24 crc kubenswrapper[4708]: E1125 05:42:24.893769 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:25 crc kubenswrapper[4708]: I1125 05:42:25.892804 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:25 crc kubenswrapper[4708]: I1125 05:42:25.892899 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:25 crc kubenswrapper[4708]: E1125 05:42:25.893007 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:25 crc kubenswrapper[4708]: E1125 05:42:25.893154 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:26 crc kubenswrapper[4708]: I1125 05:42:26.892942 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:26 crc kubenswrapper[4708]: I1125 05:42:26.893052 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:26 crc kubenswrapper[4708]: E1125 05:42:26.893103 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:26 crc kubenswrapper[4708]: E1125 05:42:26.893204 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:27 crc kubenswrapper[4708]: I1125 05:42:27.892596 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:27 crc kubenswrapper[4708]: I1125 05:42:27.892678 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:27 crc kubenswrapper[4708]: E1125 05:42:27.892734 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:27 crc kubenswrapper[4708]: E1125 05:42:27.892829 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:28 crc kubenswrapper[4708]: I1125 05:42:28.892662 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:28 crc kubenswrapper[4708]: I1125 05:42:28.892766 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:28 crc kubenswrapper[4708]: E1125 05:42:28.893405 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:28 crc kubenswrapper[4708]: E1125 05:42:28.893499 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:29 crc kubenswrapper[4708]: I1125 05:42:29.892357 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:29 crc kubenswrapper[4708]: I1125 05:42:29.892410 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:29 crc kubenswrapper[4708]: E1125 05:42:29.892476 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:29 crc kubenswrapper[4708]: E1125 05:42:29.892612 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:30 crc kubenswrapper[4708]: I1125 05:42:30.892630 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:30 crc kubenswrapper[4708]: I1125 05:42:30.892678 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:30 crc kubenswrapper[4708]: E1125 05:42:30.892753 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:30 crc kubenswrapper[4708]: E1125 05:42:30.893211 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:31 crc kubenswrapper[4708]: I1125 05:42:31.893241 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:31 crc kubenswrapper[4708]: E1125 05:42:31.893343 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:31 crc kubenswrapper[4708]: I1125 05:42:31.893396 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:31 crc kubenswrapper[4708]: E1125 05:42:31.893464 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:32 crc kubenswrapper[4708]: I1125 05:42:32.893431 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:32 crc kubenswrapper[4708]: I1125 05:42:32.893448 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:32 crc kubenswrapper[4708]: E1125 05:42:32.893550 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:32 crc kubenswrapper[4708]: E1125 05:42:32.893757 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:32 crc kubenswrapper[4708]: I1125 05:42:32.903787 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 05:42:33 crc kubenswrapper[4708]: I1125 05:42:33.893051 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:33 crc kubenswrapper[4708]: I1125 05:42:33.893051 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:33 crc kubenswrapper[4708]: E1125 05:42:33.893148 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:33 crc kubenswrapper[4708]: E1125 05:42:33.893380 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:34 crc kubenswrapper[4708]: I1125 05:42:34.124136 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:34 crc kubenswrapper[4708]: E1125 05:42:34.124235 4708 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:42:34 crc kubenswrapper[4708]: E1125 05:42:34.124291 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs podName:711da2a4-b5f2-40ce-84d4-619c4e3f2522 nodeName:}" failed. No retries permitted until 2025-11-25 05:43:38.124279603 +0000 UTC m=+159.533112988 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs") pod "network-metrics-daemon-g6lks" (UID: "711da2a4-b5f2-40ce-84d4-619c4e3f2522") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 05:42:34 crc kubenswrapper[4708]: I1125 05:42:34.892208 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:34 crc kubenswrapper[4708]: I1125 05:42:34.892246 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:34 crc kubenswrapper[4708]: E1125 05:42:34.892341 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:34 crc kubenswrapper[4708]: E1125 05:42:34.892388 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:35 crc kubenswrapper[4708]: I1125 05:42:35.892229 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:35 crc kubenswrapper[4708]: I1125 05:42:35.892229 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:35 crc kubenswrapper[4708]: E1125 05:42:35.893142 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:35 crc kubenswrapper[4708]: E1125 05:42:35.893242 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:35 crc kubenswrapper[4708]: I1125 05:42:35.893270 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:42:35 crc kubenswrapper[4708]: E1125 05:42:35.893397 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qppcd_openshift-ovn-kubernetes(6d78be28-7e93-4fda-98cd-521398f9b3e4)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" Nov 25 05:42:36 crc kubenswrapper[4708]: I1125 05:42:36.893056 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:36 crc kubenswrapper[4708]: E1125 05:42:36.893682 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:36 crc kubenswrapper[4708]: I1125 05:42:36.893236 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:36 crc kubenswrapper[4708]: E1125 05:42:36.893941 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:37 crc kubenswrapper[4708]: I1125 05:42:37.893131 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:37 crc kubenswrapper[4708]: E1125 05:42:37.893227 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:37 crc kubenswrapper[4708]: I1125 05:42:37.893352 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:37 crc kubenswrapper[4708]: E1125 05:42:37.893424 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:38 crc kubenswrapper[4708]: I1125 05:42:38.892418 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:38 crc kubenswrapper[4708]: I1125 05:42:38.892485 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:38 crc kubenswrapper[4708]: E1125 05:42:38.893386 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:38 crc kubenswrapper[4708]: E1125 05:42:38.893498 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:38 crc kubenswrapper[4708]: I1125 05:42:38.911658 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=6.911644436 podStartE2EDuration="6.911644436s" podCreationTimestamp="2025-11-25 05:42:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:38.911051891 +0000 UTC m=+100.319885277" watchObservedRunningTime="2025-11-25 05:42:38.911644436 +0000 UTC m=+100.320477822" Nov 25 05:42:39 crc kubenswrapper[4708]: I1125 05:42:39.893031 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:39 crc kubenswrapper[4708]: I1125 05:42:39.893151 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:39 crc kubenswrapper[4708]: E1125 05:42:39.893260 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:39 crc kubenswrapper[4708]: E1125 05:42:39.893449 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:40 crc kubenswrapper[4708]: I1125 05:42:40.892998 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:40 crc kubenswrapper[4708]: I1125 05:42:40.893001 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:40 crc kubenswrapper[4708]: E1125 05:42:40.893254 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:40 crc kubenswrapper[4708]: E1125 05:42:40.893131 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:41 crc kubenswrapper[4708]: I1125 05:42:41.892261 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:41 crc kubenswrapper[4708]: E1125 05:42:41.892376 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:41 crc kubenswrapper[4708]: I1125 05:42:41.892438 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:41 crc kubenswrapper[4708]: E1125 05:42:41.892485 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:42 crc kubenswrapper[4708]: I1125 05:42:42.892923 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:42 crc kubenswrapper[4708]: I1125 05:42:42.892924 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:42 crc kubenswrapper[4708]: E1125 05:42:42.893034 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:42 crc kubenswrapper[4708]: E1125 05:42:42.893100 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:43 crc kubenswrapper[4708]: I1125 05:42:43.893256 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:43 crc kubenswrapper[4708]: E1125 05:42:43.893336 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:43 crc kubenswrapper[4708]: I1125 05:42:43.893265 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:43 crc kubenswrapper[4708]: E1125 05:42:43.893402 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:44 crc kubenswrapper[4708]: I1125 05:42:44.892771 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:44 crc kubenswrapper[4708]: E1125 05:42:44.892877 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:44 crc kubenswrapper[4708]: I1125 05:42:44.892786 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:44 crc kubenswrapper[4708]: E1125 05:42:44.893184 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:45 crc kubenswrapper[4708]: I1125 05:42:45.892249 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:45 crc kubenswrapper[4708]: I1125 05:42:45.892290 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:45 crc kubenswrapper[4708]: E1125 05:42:45.892336 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:45 crc kubenswrapper[4708]: E1125 05:42:45.892397 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:46 crc kubenswrapper[4708]: I1125 05:42:46.893075 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:46 crc kubenswrapper[4708]: I1125 05:42:46.893430 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:46 crc kubenswrapper[4708]: E1125 05:42:46.893499 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:46 crc kubenswrapper[4708]: E1125 05:42:46.893584 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:46 crc kubenswrapper[4708]: I1125 05:42:46.893623 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.326377 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/3.log" Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.328426 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerStarted","Data":"672e0618a41d664b8375415fd94c0799d59202a4ee97a3f1acfa4a3d42de6ec2"} Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.329322 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.347668 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podStartSLOduration=91.347655342 podStartE2EDuration="1m31.347655342s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:47.347082884 +0000 UTC m=+108.755916270" watchObservedRunningTime="2025-11-25 05:42:47.347655342 +0000 UTC m=+108.756488727" Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.501109 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-g6lks"] Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.501222 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:47 crc kubenswrapper[4708]: E1125 05:42:47.501330 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:47 crc kubenswrapper[4708]: I1125 05:42:47.893195 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:47 crc kubenswrapper[4708]: E1125 05:42:47.894248 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:48 crc kubenswrapper[4708]: I1125 05:42:48.892689 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:48 crc kubenswrapper[4708]: I1125 05:42:48.892748 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:48 crc kubenswrapper[4708]: E1125 05:42:48.893701 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 05:42:48 crc kubenswrapper[4708]: E1125 05:42:48.893903 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 05:42:49 crc kubenswrapper[4708]: I1125 05:42:49.892684 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:49 crc kubenswrapper[4708]: I1125 05:42:49.892795 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:49 crc kubenswrapper[4708]: E1125 05:42:49.892903 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 05:42:49 crc kubenswrapper[4708]: E1125 05:42:49.893078 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-g6lks" podUID="711da2a4-b5f2-40ce-84d4-619c4e3f2522" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.154549 4708 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.154638 4708 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.177385 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.177757 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.179134 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.183611 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.184130 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.184403 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.184548 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.187001 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qtrmv"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.191368 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.191393 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.191816 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z6ctz"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.192079 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.192439 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.192684 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.193046 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.193291 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.193470 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.194574 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.194860 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.195100 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.195185 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.195909 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q294h"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196075 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196224 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196395 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196501 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6db8s"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196597 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196714 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.196883 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197152 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197280 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197551 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197868 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197309 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197433 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197580 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197608 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197639 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197663 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.198553 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.198697 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.198843 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.198863 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.199034 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197834 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.197861 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.199321 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.200046 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qhzn8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.200450 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.201989 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.205579 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.205986 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.206195 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.206435 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.206854 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.207028 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.207864 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.207890 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.209093 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-49trj"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.209631 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.209903 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.210083 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.210221 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.210240 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.210606 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.210989 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.222449 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.223586 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.223806 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224084 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224178 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224258 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224563 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224618 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.224827 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.225004 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.225147 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.225496 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.226920 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.227485 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.227806 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.228864 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.228913 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.228964 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.229027 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.228972 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.229072 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.229115 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.229261 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.237682 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.237877 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.237973 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238027 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238079 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238318 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238317 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238408 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238707 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238944 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.238998 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239211 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239289 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239421 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239712 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239823 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.239926 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.241197 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.241817 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.241865 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.241965 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.242054 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.242658 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.244305 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.244544 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.246383 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.249315 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.249591 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.249695 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.249769 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.250187 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r2zjk"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.250494 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.250686 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.250739 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.251349 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.252690 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.253122 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.253870 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l52tr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.254247 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.254579 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256166 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256313 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256436 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256617 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256782 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.256894 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257018 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257054 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257396 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257491 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257618 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257638 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257707 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257882 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.257938 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258063 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258212 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258357 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258403 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258635 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258767 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258812 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258866 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.258881 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259069 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259140 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259152 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259512 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259695 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.259996 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.260864 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.261245 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.261590 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.261628 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z4frk"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.261786 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.262445 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.264265 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.271851 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.274602 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.276130 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.276291 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.281761 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.282036 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.285854 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.291808 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.292579 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.292860 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.293022 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.294044 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.294163 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.294869 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.296320 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q294h"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.299327 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f98lb"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.300100 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z6ctz"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.300210 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.301473 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6db8s"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.303102 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qtrmv"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.303356 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-t687t"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.303877 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.303896 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.305564 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-49trj"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.306584 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.306655 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.308012 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qhzn8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.309093 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.309962 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.310923 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.311257 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l52tr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.312940 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.315204 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.315236 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-t687t"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.316204 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.317085 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.318275 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.318968 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.319785 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.319867 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.320775 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.321722 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.322603 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.323432 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.324238 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.325893 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.327020 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.328166 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.329077 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.329852 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.330710 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z4frk"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.331605 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.333071 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.335040 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f98lb"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.336773 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-r9f4b"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.337779 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.337927 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-r9f4b"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.339161 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.359441 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.380078 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.405764 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.406938 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nkt5s"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.409093 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.418498 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nkt5s"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.443852 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.459323 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.479118 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.493744 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-vnvqz"] Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.494431 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.498912 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.519055 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.540099 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.559629 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.579829 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.599149 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.619426 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.639902 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.659360 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.680154 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.699139 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.719599 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.744978 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.759232 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.784756 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.798925 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.819004 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.839813 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.859241 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.879363 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.892849 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.893015 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.899149 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.919384 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.940053 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.959232 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.979436 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 05:42:50 crc kubenswrapper[4708]: I1125 05:42:50.999391 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.019734 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.039967 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.059261 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.079491 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.099765 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.119679 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.139509 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.159898 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.178943 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.199954 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.219638 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.239993 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.258199 4708 request.go:700] Waited for 1.004931754s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcollect-profiles-dockercfg-kzf4t&limit=500&resourceVersion=0 Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.259134 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.279279 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.299469 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.318839 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.339461 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.359255 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.380059 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.399542 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.439615 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.460052 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.479134 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.499344 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.519860 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.539424 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.558959 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.579345 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.599428 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.619591 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.639898 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.658925 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.679202 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.699682 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.719364 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.740932 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.759266 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.779550 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.799918 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.819262 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.839882 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.859343 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.879975 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.893243 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.893289 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.899591 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.919430 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.939703 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.959704 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.979311 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 05:42:51 crc kubenswrapper[4708]: I1125 05:42:51.999933 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.019877 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.039143 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.059545 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.079511 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.099748 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.118954 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.139479 4708 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.158889 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.199463 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.219614 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.239910 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.258372 4708 request.go:700] Waited for 1.365265654s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-network-diagnostics/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.259321 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.273623 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.273737 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.273817 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.273893 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-auth-proxy-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.273959 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf2vq\" (UniqueName: \"kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274033 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5sg5\" (UniqueName: \"kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274112 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-srv-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274179 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mwdc\" (UniqueName: \"kubernetes.io/projected/41359abe-343a-4909-b6bf-8bc71f24fc5e-kube-api-access-9mwdc\") pod \"migrator-59844c95c7-6gm6k\" (UID: \"41359abe-343a-4909-b6bf-8bc71f24fc5e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274248 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-encryption-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274317 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-metrics-tls\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274387 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit-dir\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274455 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274514 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/062dced0-ecfd-49fa-bfbb-5ff51768816c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274596 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274661 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/062dced0-ecfd-49fa-bfbb-5ff51768816c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274748 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274815 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274875 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-serving-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.274936 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275002 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sxgn\" (UniqueName: \"kubernetes.io/projected/ab5b6672-f79c-4659-bd34-10388984b09e-kube-api-access-9sxgn\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275066 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwbwp\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-kube-api-access-rwbwp\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275123 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275182 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-service-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275242 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275304 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275433 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275496 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a55393c-9185-4059-a3ed-f79eeb2f69bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275592 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzx7k\" (UniqueName: \"kubernetes.io/projected/047cd048-1c40-42c2-8b2f-ded382715d88-kube-api-access-pzx7k\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275671 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6cjn\" (UniqueName: \"kubernetes.io/projected/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-kube-api-access-q6cjn\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275749 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-client\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275805 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275864 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/047cd048-1c40-42c2-8b2f-ded382715d88-serving-cert\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.275927 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276002 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25xwv\" (UniqueName: \"kubernetes.io/projected/062dced0-ecfd-49fa-bfbb-5ff51768816c-kube-api-access-25xwv\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276067 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276123 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-serving-cert\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276188 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqllg\" (UniqueName: \"kubernetes.io/projected/ec56d949-050a-459d-853e-5cbcf2347699-kube-api-access-jqllg\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276253 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e20fa67e-381d-49ff-82da-aff52e45d882-proxy-tls\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276542 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276607 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhbsv\" (UniqueName: \"kubernetes.io/projected/e20fa67e-381d-49ff-82da-aff52e45d882-kube-api-access-mhbsv\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276674 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276745 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276817 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gngcd\" (UniqueName: \"kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276875 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.276945 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277010 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277082 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277151 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277210 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-image-import-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277273 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277417 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-trusted-ca\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277487 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277579 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277648 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277728 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqwrj\" (UniqueName: \"kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277794 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277858 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a55393c-9185-4059-a3ed-f79eeb2f69bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277918 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87a8de16-a4f9-4750-a6ea-3796295fa455-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.277981 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278046 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpmtw\" (UniqueName: \"kubernetes.io/projected/8a55393c-9185-4059-a3ed-f79eeb2f69bf-kube-api-access-mpmtw\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278123 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-config\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278155 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278183 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278203 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278220 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab5b6672-f79c-4659-bd34-10388984b09e-machine-approver-tls\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278252 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278269 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvwgt\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278284 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/44513b1e-599c-462b-a745-da62c364079f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278301 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278328 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278344 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfqlb\" (UniqueName: \"kubernetes.io/projected/87a8de16-a4f9-4750-a6ea-3796295fa455-kube-api-access-pfqlb\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278360 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278374 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278388 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-images\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278403 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278417 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278434 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr4f9\" (UniqueName: \"kubernetes.io/projected/ac1c761e-3ccf-4c76-8e06-8e98390da65f-kube-api-access-dr4f9\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278448 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pgjw\" (UniqueName: \"kubernetes.io/projected/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-kube-api-access-7pgjw\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278464 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278479 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcv4d\" (UniqueName: \"kubernetes.io/projected/aa6199ad-a3fd-4547-83ca-06944728979d-kube-api-access-vcv4d\") pod \"downloads-7954f5f757-qtrmv\" (UID: \"aa6199ad-a3fd-4547-83ca-06944728979d\") " pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278495 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sblx\" (UniqueName: \"kubernetes.io/projected/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-kube-api-access-8sblx\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278511 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-config\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278544 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxz25\" (UniqueName: \"kubernetes.io/projected/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-kube-api-access-dxz25\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44513b1e-599c-462b-a745-da62c364079f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278575 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec56d949-050a-459d-853e-5cbcf2347699-serving-cert\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278590 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-node-pullsecrets\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.278631 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.278872 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:52.778861542 +0000 UTC m=+114.187694927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.279019 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.299723 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.319580 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.359197 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379062 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.379147 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:52.879132043 +0000 UTC m=+114.287965430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379224 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-serving-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379253 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379271 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/062dced0-ecfd-49fa-bfbb-5ff51768816c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379293 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzgds\" (UniqueName: \"kubernetes.io/projected/1634a636-2d11-464c-9306-0123e62c65d0-kube-api-access-lzgds\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379327 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379345 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46c16b8d-15a9-4d6c-a726-f7d845043a13-proxy-tls\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379362 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sxgn\" (UniqueName: \"kubernetes.io/projected/ab5b6672-f79c-4659-bd34-10388984b09e-kube-api-access-9sxgn\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379378 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379400 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379418 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-socket-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379436 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g4n8\" (UniqueName: \"kubernetes.io/projected/c6b06363-6a06-4bb6-b112-383a4e95475e-kube-api-access-2g4n8\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379461 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-images\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379476 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-plugins-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379493 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/047cd048-1c40-42c2-8b2f-ded382715d88-serving-cert\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379509 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-registration-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379542 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379559 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25xwv\" (UniqueName: \"kubernetes.io/projected/062dced0-ecfd-49fa-bfbb-5ff51768816c-kube-api-access-25xwv\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379576 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd80abac-33e8-498f-a346-7bc0e42886c1-config\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379593 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqllg\" (UniqueName: \"kubernetes.io/projected/ec56d949-050a-459d-853e-5cbcf2347699-kube-api-access-jqllg\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379611 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379627 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhbsv\" (UniqueName: \"kubernetes.io/projected/e20fa67e-381d-49ff-82da-aff52e45d882-kube-api-access-mhbsv\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379659 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gngcd\" (UniqueName: \"kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379684 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-metrics-certs\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379699 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10e41c07-da3d-4999-8919-5a7e6102a22d-cert\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379734 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379753 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379769 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-client\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379788 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-image-import-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379803 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m66l\" (UniqueName: \"kubernetes.io/projected/4c24dca1-f55a-4893-a179-ad1ce776d138-kube-api-access-2m66l\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379822 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379837 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379852 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-config\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379867 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379882 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9gg5\" (UniqueName: \"kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379897 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379913 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac43907-bde5-455d-81a1-71264ac17d51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379929 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-mountpoint-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379937 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.379945 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqwrj\" (UniqueName: \"kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380009 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87a8de16-a4f9-4750-a6ea-3796295fa455-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380043 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a55393c-9185-4059-a3ed-f79eeb2f69bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380062 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e41814f6-cdb7-4460-bc75-618b5fff39df-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380506 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-serving-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380642 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380737 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpmtw\" (UniqueName: \"kubernetes.io/projected/8a55393c-9185-4059-a3ed-f79eeb2f69bf-kube-api-access-mpmtw\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380781 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-config\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380801 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380833 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380857 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1634a636-2d11-464c-9306-0123e62c65d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380879 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-serving-cert\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380907 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/44513b1e-599c-462b-a745-da62c364079f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380924 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-service-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380944 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37759660-fe73-4292-b49d-e7572f3838ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380960 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380960 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.380977 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a770195-b221-4ec8-bbbc-92573d42db74-metrics-tls\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381050 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfqlb\" (UniqueName: \"kubernetes.io/projected/87a8de16-a4f9-4750-a6ea-3796295fa455-kube-api-access-pfqlb\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381061 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381069 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-node-bootstrap-token\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381088 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-profile-collector-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381106 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381126 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-images\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381150 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-dir\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381168 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381185 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381202 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381217 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381232 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcv4d\" (UniqueName: \"kubernetes.io/projected/aa6199ad-a3fd-4547-83ca-06944728979d-kube-api-access-vcv4d\") pod \"downloads-7954f5f757-qtrmv\" (UID: \"aa6199ad-a3fd-4547-83ca-06944728979d\") " pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381249 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sblx\" (UniqueName: \"kubernetes.io/projected/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-kube-api-access-8sblx\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381266 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-config\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381281 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381298 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqh68\" (UniqueName: \"kubernetes.io/projected/e41814f6-cdb7-4460-bc75-618b5fff39df-kube-api-access-jqh68\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381314 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxz25\" (UniqueName: \"kubernetes.io/projected/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-kube-api-access-dxz25\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381331 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x576\" (UniqueName: \"kubernetes.io/projected/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-kube-api-access-9x576\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381182 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381251 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a55393c-9185-4059-a3ed-f79eeb2f69bf-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381476 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-encryption-config\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381494 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-node-pullsecrets\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381512 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-policies\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381546 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-cabundle\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381546 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-config\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381562 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381578 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-stats-auth\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381595 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381616 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-encryption-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381632 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf2vq\" (UniqueName: \"kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381649 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5sg5\" (UniqueName: \"kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381675 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37759660-fe73-4292-b49d-e7572f3838ee-config\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381701 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-metrics-tls\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381716 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r7cr\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-kube-api-access-9r7cr\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381736 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zstv9\" (UniqueName: \"kubernetes.io/projected/847c28aa-0945-492d-941c-b29304b7768b-kube-api-access-zstv9\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381757 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-serving-cert\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381782 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381800 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381821 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381828 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381842 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-webhook-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381949 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwbwp\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-kube-api-access-rwbwp\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381969 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-node-pullsecrets\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.381990 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382024 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhcxj\" (UniqueName: \"kubernetes.io/projected/10e41c07-da3d-4999-8919-5a7e6102a22d-kube-api-access-rhcxj\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382049 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382073 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-service-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382095 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382121 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a55393c-9185-4059-a3ed-f79eeb2f69bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382149 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382155 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ac43907-bde5-455d-81a1-71264ac17d51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382216 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdz5h\" (UniqueName: \"kubernetes.io/projected/db4f79f6-d226-40c5-bf33-e0f545525833-kube-api-access-zdz5h\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382232 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382321 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382325 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzx7k\" (UniqueName: \"kubernetes.io/projected/047cd048-1c40-42c2-8b2f-ded382715d88-kube-api-access-pzx7k\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382364 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6cjn\" (UniqueName: \"kubernetes.io/projected/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-kube-api-access-q6cjn\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382386 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382406 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-client\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382422 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382440 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382457 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-serving-cert\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382474 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37759660-fe73-4292-b49d-e7572f3838ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382490 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382540 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e20fa67e-381d-49ff-82da-aff52e45d882-proxy-tls\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382559 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382578 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-csi-data-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382711 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-auth-proxy-config\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.382724 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ec56d949-050a-459d-853e-5cbcf2347699-service-ca-bundle\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.383070 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.383197 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384235 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/87a8de16-a4f9-4750-a6ea-3796295fa455-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384445 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384887 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b06363-6a06-4bb6-b112-383a4e95475e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384921 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nsbp\" (UniqueName: \"kubernetes.io/projected/76500481-c48a-4cf4-993e-ba1f81dc4327-kube-api-access-6nsbp\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384932 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384947 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384971 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1634a636-2d11-464c-9306-0123e62c65d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.384984 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385031 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385040 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/062dced0-ecfd-49fa-bfbb-5ff51768816c-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385053 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385073 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxkgw\" (UniqueName: \"kubernetes.io/projected/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-kube-api-access-dxkgw\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385241 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd80abac-33e8-498f-a346-7bc0e42886c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385340 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cr7z\" (UniqueName: \"kubernetes.io/projected/9edd4399-bff5-4125-be89-3519a2d94ab4-kube-api-access-2cr7z\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385362 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a770195-b221-4ec8-bbbc-92573d42db74-config-volume\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385419 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385476 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385503 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xf7hz\" (UniqueName: \"kubernetes.io/projected/389881b2-1179-4817-8bd0-9628ddbdaf31-kube-api-access-xf7hz\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385568 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385599 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b06363-6a06-4bb6-b112-383a4e95475e-serving-cert\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385627 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.386814 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-image-import-ca\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.386957 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.386965 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387295 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-trusted-ca-bundle\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387454 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387545 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.385632 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-trusted-ca\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387626 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnlkw\" (UniqueName: \"kubernetes.io/projected/46c16b8d-15a9-4d6c-a726-f7d845043a13-kube-api-access-cnlkw\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387649 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9edd4399-bff5-4125-be89-3519a2d94ab4-tmpfs\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387686 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387693 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387700 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a55393c-9185-4059-a3ed-f79eeb2f69bf-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387714 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46c16b8d-15a9-4d6c-a726-f7d845043a13-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387735 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e20fa67e-381d-49ff-82da-aff52e45d882-images\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387747 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387782 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4f79f6-d226-40c5-bf33-e0f545525833-service-ca-bundle\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387802 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387804 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387928 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdnvg\" (UniqueName: \"kubernetes.io/projected/7c9f0208-cc82-49bc-907c-7f9f81832fa7-kube-api-access-pdnvg\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387954 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/847c28aa-0945-492d-941c-b29304b7768b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.387989 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388014 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab5b6672-f79c-4659-bd34-10388984b09e-machine-approver-tls\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388057 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388083 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-certs\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388106 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-serving-cert\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388129 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvwgt\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388257 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388297 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-key\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388324 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388346 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-client\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.388367 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:52.888357197 +0000 UTC m=+114.297190584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388398 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388428 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr4f9\" (UniqueName: \"kubernetes.io/projected/ac1c761e-3ccf-4c76-8e06-8e98390da65f-kube-api-access-dr4f9\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388459 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pgjw\" (UniqueName: \"kubernetes.io/projected/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-kube-api-access-7pgjw\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388603 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzbrd\" (UniqueName: \"kubernetes.io/projected/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-kube-api-access-dzbrd\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388633 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44513b1e-599c-462b-a745-da62c364079f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388654 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-srv-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388690 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-config\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388713 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec56d949-050a-459d-853e-5cbcf2347699-serving-cert\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388756 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd80abac-33e8-498f-a346-7bc0e42886c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388784 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkhvx\" (UniqueName: \"kubernetes.io/projected/9a770195-b221-4ec8-bbbc-92573d42db74-kube-api-access-zkhvx\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388788 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-trusted-ca\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388804 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388833 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388863 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388887 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388904 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388931 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-auth-proxy-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388955 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-srv-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388981 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mwdc\" (UniqueName: \"kubernetes.io/projected/41359abe-343a-4909-b6bf-8bc71f24fc5e-kube-api-access-9mwdc\") pod \"migrator-59844c95c7-6gm6k\" (UID: \"41359abe-343a-4909-b6bf-8bc71f24fc5e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389002 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit-dir\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389031 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389055 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/062dced0-ecfd-49fa-bfbb-5ff51768816c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389095 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-default-certificate\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389119 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgs6c\" (UniqueName: \"kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389140 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-config\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.389876 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.390348 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.390421 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.390799 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/44513b1e-599c-462b-a745-da62c364079f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.390949 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.388602 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-etcd-client\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391025 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44513b1e-599c-462b-a745-da62c364079f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391277 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391338 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391487 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ab5b6672-f79c-4659-bd34-10388984b09e-auth-proxy-config\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391584 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-encryption-config\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.391593 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/047cd048-1c40-42c2-8b2f-ded382715d88-config\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.392339 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.392611 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.392671 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-audit-dir\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.392938 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.392975 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/062dced0-ecfd-49fa-bfbb-5ff51768816c-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.394678 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-metrics-tls\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.395595 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396204 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396278 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396303 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/047cd048-1c40-42c2-8b2f-ded382715d88-serving-cert\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396339 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-serving-cert\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396884 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e20fa67e-381d-49ff-82da-aff52e45d882-proxy-tls\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.396966 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ac1c761e-3ccf-4c76-8e06-8e98390da65f-srv-cert\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.397031 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.398104 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.398138 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ec56d949-050a-459d-853e-5cbcf2347699-serving-cert\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.398193 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ab5b6672-f79c-4659-bd34-10388984b09e-machine-approver-tls\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.398613 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.431220 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sxgn\" (UniqueName: \"kubernetes.io/projected/ab5b6672-f79c-4659-bd34-10388984b09e-kube-api-access-9sxgn\") pod \"machine-approver-56656f9798-6vkgb\" (UID: \"ab5b6672-f79c-4659-bd34-10388984b09e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.450484 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.452208 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.470755 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25xwv\" (UniqueName: \"kubernetes.io/projected/062dced0-ecfd-49fa-bfbb-5ff51768816c-kube-api-access-25xwv\") pod \"kube-storage-version-migrator-operator-b67b599dd-rzbf4\" (UID: \"062dced0-ecfd-49fa-bfbb-5ff51768816c\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489789 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489902 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-srv-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489931 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd80abac-33e8-498f-a346-7bc0e42886c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489953 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489969 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkhvx\" (UniqueName: \"kubernetes.io/projected/9a770195-b221-4ec8-bbbc-92573d42db74-kube-api-access-zkhvx\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.489989 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490015 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-default-certificate\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490031 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgs6c\" (UniqueName: \"kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490049 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-config\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490066 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzgds\" (UniqueName: \"kubernetes.io/projected/1634a636-2d11-464c-9306-0123e62c65d0-kube-api-access-lzgds\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490086 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46c16b8d-15a9-4d6c-a726-f7d845043a13-proxy-tls\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490111 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-socket-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490242 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-socket-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.490368 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:52.990347545 +0000 UTC m=+114.399180931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490456 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g4n8\" (UniqueName: \"kubernetes.io/projected/c6b06363-6a06-4bb6-b112-383a4e95475e-kube-api-access-2g4n8\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490567 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-images\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490644 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-plugins-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490718 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-registration-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490779 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-registration-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490720 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-plugins-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490825 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490791 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd80abac-33e8-498f-a346-7bc0e42886c1-config\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.490978 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-metrics-certs\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491048 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10e41c07-da3d-4999-8919-5a7e6102a22d-cert\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491116 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491185 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-client\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491156 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-config\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491255 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m66l\" (UniqueName: \"kubernetes.io/projected/4c24dca1-f55a-4893-a179-ad1ce776d138-kube-api-access-2m66l\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491317 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-config\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491334 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-images\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491347 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9gg5\" (UniqueName: \"kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491388 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac43907-bde5-455d-81a1-71264ac17d51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491410 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-mountpoint-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491438 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e41814f6-cdb7-4460-bc75-618b5fff39df-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491469 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1634a636-2d11-464c-9306-0123e62c65d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491487 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-serving-cert\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491541 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-service-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491565 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37759660-fe73-4292-b49d-e7572f3838ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491582 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491591 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fd80abac-33e8-498f-a346-7bc0e42886c1-config\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491601 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a770195-b221-4ec8-bbbc-92573d42db74-metrics-tls\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491653 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-node-bootstrap-token\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-profile-collector-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491712 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-dir\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491755 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-config\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491773 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491791 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqh68\" (UniqueName: \"kubernetes.io/projected/e41814f6-cdb7-4460-bc75-618b5fff39df-kube-api-access-jqh68\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491832 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-encryption-config\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491851 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x576\" (UniqueName: \"kubernetes.io/projected/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-kube-api-access-9x576\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491871 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-policies\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491886 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-cabundle\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491904 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-stats-auth\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491920 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491963 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37759660-fe73-4292-b49d-e7572f3838ee-config\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.491981 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r7cr\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-kube-api-access-9r7cr\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492000 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zstv9\" (UniqueName: \"kubernetes.io/projected/847c28aa-0945-492d-941c-b29304b7768b-kube-api-access-zstv9\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492014 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-serving-cert\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492029 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492050 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-webhook-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492067 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhcxj\" (UniqueName: \"kubernetes.io/projected/10e41c07-da3d-4999-8919-5a7e6102a22d-kube-api-access-rhcxj\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492079 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492093 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ac43907-bde5-455d-81a1-71264ac17d51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492110 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdz5h\" (UniqueName: \"kubernetes.io/projected/db4f79f6-d226-40c5-bf33-e0f545525833-kube-api-access-zdz5h\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492113 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqllg\" (UniqueName: \"kubernetes.io/projected/ec56d949-050a-459d-853e-5cbcf2347699-kube-api-access-jqllg\") pod \"authentication-operator-69f744f599-6db8s\" (UID: \"ec56d949-050a-459d-853e-5cbcf2347699\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492132 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492151 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37759660-fe73-4292-b49d-e7572f3838ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492168 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492195 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492211 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-csi-data-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492227 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b06363-6a06-4bb6-b112-383a4e95475e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492244 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nsbp\" (UniqueName: \"kubernetes.io/projected/76500481-c48a-4cf4-993e-ba1f81dc4327-kube-api-access-6nsbp\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492266 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1634a636-2d11-464c-9306-0123e62c65d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492291 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxkgw\" (UniqueName: \"kubernetes.io/projected/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-kube-api-access-dxkgw\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492311 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd80abac-33e8-498f-a346-7bc0e42886c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492329 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cr7z\" (UniqueName: \"kubernetes.io/projected/9edd4399-bff5-4125-be89-3519a2d94ab4-kube-api-access-2cr7z\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492345 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a770195-b221-4ec8-bbbc-92573d42db74-config-volume\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492370 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xf7hz\" (UniqueName: \"kubernetes.io/projected/389881b2-1179-4817-8bd0-9628ddbdaf31-kube-api-access-xf7hz\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492389 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492405 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b06363-6a06-4bb6-b112-383a4e95475e-serving-cert\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492421 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492439 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46c16b8d-15a9-4d6c-a726-f7d845043a13-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492458 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnlkw\" (UniqueName: \"kubernetes.io/projected/46c16b8d-15a9-4d6c-a726-f7d845043a13-kube-api-access-cnlkw\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492476 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9edd4399-bff5-4125-be89-3519a2d94ab4-tmpfs\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492499 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4f79f6-d226-40c5-bf33-e0f545525833-service-ca-bundle\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492688 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/46c16b8d-15a9-4d6c-a726-f7d845043a13-proxy-tls\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/847c28aa-0945-492d-941c-b29304b7768b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492739 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdnvg\" (UniqueName: \"kubernetes.io/projected/7c9f0208-cc82-49bc-907c-7f9f81832fa7-kube-api-access-pdnvg\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492764 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492792 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-certs\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492809 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-serving-cert\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492839 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-key\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492850 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492858 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-client\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492916 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492936 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-cabundle\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492951 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzbrd\" (UniqueName: \"kubernetes.io/projected/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-kube-api-access-dzbrd\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.493139 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-dir\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.493347 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-mountpoint-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.493614 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-etcd-client\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.493696 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9a770195-b221-4ec8-bbbc-92573d42db74-metrics-tls\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494260 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-config\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494704 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/10e41c07-da3d-4999-8919-5a7e6102a22d-cert\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494737 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-service-ca\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494757 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-etcd-client\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494855 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.494871 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-profile-collector-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.495050 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:52.995042527 +0000 UTC m=+114.403875914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.495230 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ac43907-bde5-455d-81a1-71264ac17d51-trusted-ca\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.495837 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-default-certificate\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.495933 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.496077 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-encryption-config\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.496164 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37759660-fe73-4292-b49d-e7572f3838ee-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.496405 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/847c28aa-0945-492d-941c-b29304b7768b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.496485 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-csi-data-dir\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.497350 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/76500481-c48a-4cf4-993e-ba1f81dc4327-srv-cert\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.497543 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-audit-policies\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.497660 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-serving-cert\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.497938 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-metrics-certs\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.498208 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.498316 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c6b06363-6a06-4bb6-b112-383a4e95475e-available-featuregates\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.498372 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1634a636-2d11-464c-9306-0123e62c65d0-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.498782 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1634a636-2d11-464c-9306-0123e62c65d0-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.498845 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9edd4399-bff5-4125-be89-3519a2d94ab4-tmpfs\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.492091 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e41814f6-cdb7-4460-bc75-618b5fff39df-config\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.499557 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-certs\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.499758 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/db4f79f6-d226-40c5-bf33-e0f545525833-stats-auth\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.500033 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-webhook-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.500253 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/389881b2-1179-4817-8bd0-9628ddbdaf31-node-bootstrap-token\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.500586 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.500682 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/e41814f6-cdb7-4460-bc75-618b5fff39df-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.500694 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501119 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9ac43907-bde5-455d-81a1-71264ac17d51-metrics-tls\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501280 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4f79f6-d226-40c5-bf33-e0f545525833-service-ca-bundle\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501473 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd80abac-33e8-498f-a346-7bc0e42886c1-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501475 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/46c16b8d-15a9-4d6c-a726-f7d845043a13-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501477 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501552 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37759660-fe73-4292-b49d-e7572f3838ee-config\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501889 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c9f0208-cc82-49bc-907c-7f9f81832fa7-serving-cert\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.501947 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9a770195-b221-4ec8-bbbc-92573d42db74-config-volume\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.502656 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9edd4399-bff5-4125-be89-3519a2d94ab4-apiservice-cert\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.503429 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-serving-cert\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.503676 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.504020 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.504139 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4c24dca1-f55a-4893-a179-ad1ce776d138-signing-key\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.505094 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b06363-6a06-4bb6-b112-383a4e95475e-serving-cert\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.509966 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gngcd\" (UniqueName: \"kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd\") pod \"marketplace-operator-79b997595-mr2tz\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.527022 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.530359 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sblx\" (UniqueName: \"kubernetes.io/projected/6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67-kube-api-access-8sblx\") pod \"cluster-samples-operator-665b6dd947-97g7p\" (UID: \"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.553180 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfqlb\" (UniqueName: \"kubernetes.io/projected/87a8de16-a4f9-4750-a6ea-3796295fa455-kube-api-access-pfqlb\") pod \"multus-admission-controller-857f4d67dd-49trj\" (UID: \"87a8de16-a4f9-4750-a6ea-3796295fa455\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.573049 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxz25\" (UniqueName: \"kubernetes.io/projected/e1dcf9d6-33d7-4482-b24f-990de83f2fa1-kube-api-access-dxz25\") pod \"control-plane-machine-set-operator-78cbb6b69f-642cw\" (UID: \"e1dcf9d6-33d7-4482-b24f-990de83f2fa1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.591644 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzx7k\" (UniqueName: \"kubernetes.io/projected/047cd048-1c40-42c2-8b2f-ded382715d88-kube-api-access-pzx7k\") pod \"console-operator-58897d9998-z6ctz\" (UID: \"047cd048-1c40-42c2-8b2f-ded382715d88\") " pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.593595 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.593780 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.093756311 +0000 UTC m=+114.502589697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.594082 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.594467 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.094459583 +0000 UTC m=+114.503292970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.613794 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwbwp\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-kube-api-access-rwbwp\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.632290 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6cjn\" (UniqueName: \"kubernetes.io/projected/f9d144cb-ff11-40a4-a30a-d4feb8c3d58c-kube-api-access-q6cjn\") pod \"dns-operator-744455d44c-q294h\" (UID: \"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c\") " pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.641164 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.652808 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhbsv\" (UniqueName: \"kubernetes.io/projected/e20fa67e-381d-49ff-82da-aff52e45d882-kube-api-access-mhbsv\") pod \"machine-config-operator-74547568cd-4wdjr\" (UID: \"e20fa67e-381d-49ff-82da-aff52e45d882\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.665854 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.671654 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqwrj\" (UniqueName: \"kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj\") pod \"controller-manager-879f6c89f-6gs2b\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.693423 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf2vq\" (UniqueName: \"kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq\") pod \"oauth-openshift-558db77b4-2kwv6\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.694939 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.695091 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.19506753 +0000 UTC m=+114.603900917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.695227 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.695568 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.195558103 +0000 UTC m=+114.604391489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.699358 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.711838 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpmtw\" (UniqueName: \"kubernetes.io/projected/8a55393c-9185-4059-a3ed-f79eeb2f69bf-kube-api-access-mpmtw\") pod \"openshift-controller-manager-operator-756b6f6bc6-5tpvl\" (UID: \"8a55393c-9185-4059-a3ed-f79eeb2f69bf\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.712335 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.718895 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.723821 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.732625 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5sg5\" (UniqueName: \"kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5\") pod \"route-controller-manager-6576b87f9c-gp6rv\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.743586 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.753790 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e29a355-332e-4f3b-a31a-20f5a0eaaaad-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-n58zc\" (UID: \"7e29a355-332e-4f3b-a31a-20f5a0eaaaad\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.757588 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.768769 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.773805 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcv4d\" (UniqueName: \"kubernetes.io/projected/aa6199ad-a3fd-4547-83ca-06944728979d-kube-api-access-vcv4d\") pod \"downloads-7954f5f757-qtrmv\" (UID: \"aa6199ad-a3fd-4547-83ca-06944728979d\") " pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.779587 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z6ctz"] Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.796562 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.797338 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.297321733 +0000 UTC m=+114.706155120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.805642 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/44513b1e-599c-462b-a745-da62c364079f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-snrvr\" (UID: \"44513b1e-599c-462b-a745-da62c364079f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.805814 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.812382 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.814397 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvwgt\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: W1125 05:42:52.832072 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod047cd048_1c40_42c2_8b2f_ded382715d88.slice/crio-0fd345a2fcdd8209736c78fbb58de3173620f5338461e38a0a68567d51bb2c94 WatchSource:0}: Error finding container 0fd345a2fcdd8209736c78fbb58de3173620f5338461e38a0a68567d51bb2c94: Status 404 returned error can't find the container with id 0fd345a2fcdd8209736c78fbb58de3173620f5338461e38a0a68567d51bb2c94 Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.832944 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pgjw\" (UniqueName: \"kubernetes.io/projected/c743ab1a-6a44-49e0-b9ec-ccf8b8213a21-kube-api-access-7pgjw\") pod \"apiserver-76f77b778f-qhzn8\" (UID: \"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21\") " pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.858734 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr4f9\" (UniqueName: \"kubernetes.io/projected/ac1c761e-3ccf-4c76-8e06-8e98390da65f-kube-api-access-dr4f9\") pod \"olm-operator-6b444d44fb-n578c\" (UID: \"ac1c761e-3ccf-4c76-8e06-8e98390da65f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.880174 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mwdc\" (UniqueName: \"kubernetes.io/projected/41359abe-343a-4909-b6bf-8bc71f24fc5e-kube-api-access-9mwdc\") pod \"migrator-59844c95c7-6gm6k\" (UID: \"41359abe-343a-4909-b6bf-8bc71f24fc5e\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.898992 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:52 crc kubenswrapper[4708]: E1125 05:42:52.899974 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.399959148 +0000 UTC m=+114.808792534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.902485 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.911959 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.917623 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzgds\" (UniqueName: \"kubernetes.io/projected/1634a636-2d11-464c-9306-0123e62c65d0-kube-api-access-lzgds\") pod \"openshift-apiserver-operator-796bbdcf4f-2fgg4\" (UID: \"1634a636-2d11-464c-9306-0123e62c65d0\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.937978 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/fd80abac-33e8-498f-a346-7bc0e42886c1-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-sskm9\" (UID: \"fd80abac-33e8-498f-a346-7bc0e42886c1\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.952979 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.959400 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgs6c\" (UniqueName: \"kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c\") pod \"console-f9d7485db-zkmtb\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.975007 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkhvx\" (UniqueName: \"kubernetes.io/projected/9a770195-b221-4ec8-bbbc-92573d42db74-kube-api-access-zkhvx\") pod \"dns-default-r9f4b\" (UID: \"9a770195-b221-4ec8-bbbc-92573d42db74\") " pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.988120 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.988789 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" Nov 25 05:42:52 crc kubenswrapper[4708]: I1125 05:42:52.994854 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-bound-sa-token\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.001158 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.002694 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.502669309 +0000 UTC m=+114.911502694 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.003680 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.004053 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.504040358 +0000 UTC m=+114.912873744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.013312 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g4n8\" (UniqueName: \"kubernetes.io/projected/c6b06363-6a06-4bb6-b112-383a4e95475e-kube-api-access-2g4n8\") pod \"openshift-config-operator-7777fb866f-gqbkc\" (UID: \"c6b06363-6a06-4bb6-b112-383a4e95475e\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.029506 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.033305 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.036239 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9gg5\" (UniqueName: \"kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5\") pod \"collect-profiles-29400810-qcn4r\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.057356 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-49trj"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.059292 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqh68\" (UniqueName: \"kubernetes.io/projected/e41814f6-cdb7-4460-bc75-618b5fff39df-kube-api-access-jqh68\") pod \"machine-api-operator-5694c8668f-f98lb\" (UID: \"e41814f6-cdb7-4460-bc75-618b5fff39df\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.069887 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.073807 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.076180 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzbrd\" (UniqueName: \"kubernetes.io/projected/0cfd3a04-2b25-4da4-a565-69b909c2d5e3-kube-api-access-dzbrd\") pod \"apiserver-7bbb656c7d-kgp2r\" (UID: \"0cfd3a04-2b25-4da4-a565-69b909c2d5e3\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.093019 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m66l\" (UniqueName: \"kubernetes.io/projected/4c24dca1-f55a-4893-a179-ad1ce776d138-kube-api-access-2m66l\") pod \"service-ca-9c57cc56f-l52tr\" (UID: \"4c24dca1-f55a-4893-a179-ad1ce776d138\") " pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.106161 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.116472 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.616443694 +0000 UTC m=+115.025277081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.116777 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.117773 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.617758528 +0000 UTC m=+115.026591914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.121165 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.125159 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdnvg\" (UniqueName: \"kubernetes.io/projected/7c9f0208-cc82-49bc-907c-7f9f81832fa7-kube-api-access-pdnvg\") pod \"etcd-operator-b45778765-z4frk\" (UID: \"7c9f0208-cc82-49bc-907c-7f9f81832fa7\") " pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.133733 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q294h"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.136900 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.137613 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x576\" (UniqueName: \"kubernetes.io/projected/f3e8be91-d1a1-4107-bb4c-67a04cfd93cb-kube-api-access-9x576\") pod \"service-ca-operator-777779d784-kgmw8\" (UID: \"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.140909 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.158101 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.168714 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.170489 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r7cr\" (UniqueName: \"kubernetes.io/projected/9ac43907-bde5-455d-81a1-71264ac17d51-kube-api-access-9r7cr\") pod \"ingress-operator-5b745b69d9-nlcx6\" (UID: \"9ac43907-bde5-455d-81a1-71264ac17d51\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.174378 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.174595 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" Nov 25 05:42:53 crc kubenswrapper[4708]: W1125 05:42:53.176269 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9d144cb_ff11_40a4_a30a_d4feb8c3d58c.slice/crio-9942fe3e6d5294db079b125a67ab4f06aa3f550927633738d9bd6c6bf45981b2 WatchSource:0}: Error finding container 9942fe3e6d5294db079b125a67ab4f06aa3f550927633738d9bd6c6bf45981b2: Status 404 returned error can't find the container with id 9942fe3e6d5294db079b125a67ab4f06aa3f550927633738d9bd6c6bf45981b2 Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.179124 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.186315 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.188599 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-6db8s"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.189627 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qtrmv"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.191496 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xf7hz\" (UniqueName: \"kubernetes.io/projected/389881b2-1179-4817-8bd0-9628ddbdaf31-kube-api-access-xf7hz\") pod \"machine-config-server-vnvqz\" (UID: \"389881b2-1179-4817-8bd0-9628ddbdaf31\") " pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.196142 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nsbp\" (UniqueName: \"kubernetes.io/projected/76500481-c48a-4cf4-993e-ba1f81dc4327-kube-api-access-6nsbp\") pod \"catalog-operator-68c6474976-schhr\" (UID: \"76500481-c48a-4cf4-993e-ba1f81dc4327\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.199890 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.214376 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdz5h\" (UniqueName: \"kubernetes.io/projected/db4f79f6-d226-40c5-bf33-e0f545525833-kube-api-access-zdz5h\") pod \"router-default-5444994796-r2zjk\" (UID: \"db4f79f6-d226-40c5-bf33-e0f545525833\") " pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.215106 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.218071 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.218507 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.718492121 +0000 UTC m=+115.127325506 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.221682 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.232667 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.233091 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhcxj\" (UniqueName: \"kubernetes.io/projected/10e41c07-da3d-4999-8919-5a7e6102a22d-kube-api-access-rhcxj\") pod \"ingress-canary-t687t\" (UID: \"10e41c07-da3d-4999-8919-5a7e6102a22d\") " pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.261314 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.261884 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-vnvqz" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.262412 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.270261 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnlkw\" (UniqueName: \"kubernetes.io/projected/46c16b8d-15a9-4d6c-a726-f7d845043a13-kube-api-access-cnlkw\") pod \"machine-config-controller-84d6567774-kcnvk\" (UID: \"46c16b8d-15a9-4d6c-a726-f7d845043a13\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.274224 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxkgw\" (UniqueName: \"kubernetes.io/projected/392e81c8-e2ab-4bd2-97b1-ba0553e6b485-kube-api-access-dxkgw\") pod \"csi-hostpathplugin-nkt5s\" (UID: \"392e81c8-e2ab-4bd2-97b1-ba0553e6b485\") " pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.300067 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cr7z\" (UniqueName: \"kubernetes.io/projected/9edd4399-bff5-4125-be89-3519a2d94ab4-kube-api-access-2cr7z\") pod \"packageserver-d55dfcdfc-vjzpn\" (UID: \"9edd4399-bff5-4125-be89-3519a2d94ab4\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:53 crc kubenswrapper[4708]: W1125 05:42:53.307705 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f7427b2_e742_48fb_8f6f_f6f6b6c1cd5c.slice/crio-820cf8151b677b0081ef1324a393dd5bca3e3725c3638abfe287cff77d9abce6 WatchSource:0}: Error finding container 820cf8151b677b0081ef1324a393dd5bca3e3725c3638abfe287cff77d9abce6: Status 404 returned error can't find the container with id 820cf8151b677b0081ef1324a393dd5bca3e3725c3638abfe287cff77d9abce6 Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.315950 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/37759660-fe73-4292-b49d-e7572f3838ee-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-mhxgn\" (UID: \"37759660-fe73-4292-b49d-e7572f3838ee\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.320070 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.320416 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.820404661 +0000 UTC m=+115.229238047 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.323040 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.342876 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zstv9\" (UniqueName: \"kubernetes.io/projected/847c28aa-0945-492d-941c-b29304b7768b-kube-api-access-zstv9\") pod \"package-server-manager-789f6589d5-wd8m8\" (UID: \"847c28aa-0945-492d-941c-b29304b7768b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.382232 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" event={"ID":"ec56d949-050a-459d-853e-5cbcf2347699","Type":"ContainerStarted","Data":"9cf420551d3380e8da8ac88dce50985a2b3943886aa2d214b91e3bac1959f851"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.384088 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qtrmv" event={"ID":"aa6199ad-a3fd-4547-83ca-06944728979d","Type":"ContainerStarted","Data":"9b2c57d86ce63b926c0ff7544550af2311ed28fd19b26f2f3daa02ec134a46a6"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.389496 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" event={"ID":"062dced0-ecfd-49fa-bfbb-5ff51768816c","Type":"ContainerStarted","Data":"26ce51394478913885edf0c88bceecdfbf2d31393586e229fe82e42fc73e515d"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.391587 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" event={"ID":"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c","Type":"ContainerStarted","Data":"9942fe3e6d5294db079b125a67ab4f06aa3f550927633738d9bd6c6bf45981b2"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.396094 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" event={"ID":"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c","Type":"ContainerStarted","Data":"820cf8151b677b0081ef1324a393dd5bca3e3725c3638abfe287cff77d9abce6"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.397790 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" event={"ID":"e20fa67e-381d-49ff-82da-aff52e45d882","Type":"ContainerStarted","Data":"b7f9e7071b5f4303a73ec5f691342653ed4b00997e306b4f3541e71d48a342b7"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.399135 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" event={"ID":"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67","Type":"ContainerStarted","Data":"ae8863a0b022a6c8980ba94cb889391769043708da2bdf81299500e62f58fd33"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.400190 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" event={"ID":"44513b1e-599c-462b-a745-da62c364079f","Type":"ContainerStarted","Data":"c08a1408cb58c3684a54a76a8245a2471cdd4f7619d06d92be0fa66c7043191c"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.400218 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" event={"ID":"44513b1e-599c-462b-a745-da62c364079f","Type":"ContainerStarted","Data":"7cacb9efc165c4b6d6f212369931d9cb4416ce0562dee6193cef87e99d40251b"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.407896 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.408445 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" event={"ID":"ab5b6672-f79c-4659-bd34-10388984b09e","Type":"ContainerStarted","Data":"807b62d1574a55b12a31f940e643177e68c5d26eb3a9626d4df7b5aff4d11423"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.408470 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" event={"ID":"ab5b6672-f79c-4659-bd34-10388984b09e","Type":"ContainerStarted","Data":"16ce0d63a782335c3d9c31c870a1aa52a301138dc7e26262b35ad17ed1f14e10"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.408480 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" event={"ID":"ab5b6672-f79c-4659-bd34-10388984b09e","Type":"ContainerStarted","Data":"3d1832516787ef87a3e18441caf77b2fdfd521da57a3e73e0b00380af4a01a39"} Nov 25 05:42:53 crc kubenswrapper[4708]: W1125 05:42:53.418047 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1dcf9d6_33d7_4482_b24f_990de83f2fa1.slice/crio-0e9b98a6d2befdb72f5f4b8fcc82d36e47029b75c2dcbc44e8bb3c1aef2b2466 WatchSource:0}: Error finding container 0e9b98a6d2befdb72f5f4b8fcc82d36e47029b75c2dcbc44e8bb3c1aef2b2466: Status 404 returned error can't find the container with id 0e9b98a6d2befdb72f5f4b8fcc82d36e47029b75c2dcbc44e8bb3c1aef2b2466 Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.418130 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.422059 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.422340 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:53.922327622 +0000 UTC m=+115.331161007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.423770 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" event={"ID":"047cd048-1c40-42c2-8b2f-ded382715d88","Type":"ContainerStarted","Data":"5a76b4b83a54f30bf62b81deb4c1bdfdd689a76fe328712c506c553745ab2f4a"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.423790 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" event={"ID":"047cd048-1c40-42c2-8b2f-ded382715d88","Type":"ContainerStarted","Data":"0fd345a2fcdd8209736c78fbb58de3173620f5338461e38a0a68567d51bb2c94"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.424356 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.426012 4708 patch_prober.go:28] interesting pod/console-operator-58897d9998-z6ctz container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.426038 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" podUID="047cd048-1c40-42c2-8b2f-ded382715d88" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.6:8443/readyz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.437685 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" event={"ID":"115b7e0c-95db-4ead-8f17-d1b9941bc933","Type":"ContainerStarted","Data":"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.437719 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" event={"ID":"115b7e0c-95db-4ead-8f17-d1b9941bc933","Type":"ContainerStarted","Data":"44a48f8bfd99cc5c30c395ae25e09602aa7932e6295156c563effea38271ce96"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.438041 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.442290 4708 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-mr2tz container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.442322 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.447817 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.451606 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.463815 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.477218 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" event={"ID":"87a8de16-a4f9-4750-a6ea-3796295fa455","Type":"ContainerStarted","Data":"71ca1292e02a05112b1288a198e751e6aa6aa42b19cd4ff92bd35c03b82d88b1"} Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.480351 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.489679 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.506686 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.509126 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.521852 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.527132 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.529073 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.02906029 +0000 UTC m=+115.437893676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.530295 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.531098 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-t687t" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.531223 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.561420 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.629386 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.629813 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.129798231 +0000 UTC m=+115.538631617 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.654587 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-qhzn8"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.668105 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.740243 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.740785 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.240729558 +0000 UTC m=+115.649562944 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.818713 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.820197 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.841019 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.841282 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.341268454 +0000 UTC m=+115.750101841 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:53 crc kubenswrapper[4708]: I1125 05:42:53.949251 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:53 crc kubenswrapper[4708]: E1125 05:42:53.949680 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.44965782 +0000 UTC m=+115.858491206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.029107 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.047448 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.050600 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.051422 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.551404438 +0000 UTC m=+115.960237825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.080166 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-l52tr"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.082560 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.085629 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-f98lb"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.156423 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.160935 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.660917287 +0000 UTC m=+116.069750673 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: W1125 05:42:54.219572 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc35d3991_35e6_4cd7_a3c9_11513018d1fd.slice/crio-1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd WatchSource:0}: Error finding container 1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd: Status 404 returned error can't find the container with id 1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.257241 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.257603 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.757584901 +0000 UTC m=+116.166418287 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.257991 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.258251 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.758243549 +0000 UTC m=+116.167076936 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.270403 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.321813 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-z4frk"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.361968 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.363177 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.86315953 +0000 UTC m=+116.271992916 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.368566 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.387871 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-r9f4b"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.402779 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.466118 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.466138 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.466601 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:54.966585328 +0000 UTC m=+116.375418714 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.470128 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.477226 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-nkt5s"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.480798 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.491350 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkmtb" event={"ID":"c3776f93-7ce9-4056-86c8-7a2d95a02650","Type":"ContainerStarted","Data":"030fd166ce25a6d774455f427f1f904879725265278f754454cbe09f4f876fa7"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.496200 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" event={"ID":"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c","Type":"ContainerStarted","Data":"b2c691b9cead19e83cd04aa4e81fe8a34168a626d2d2fce20b05cbd3b7897efe"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.500182 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" event={"ID":"46c16b8d-15a9-4d6c-a726-f7d845043a13","Type":"ContainerStarted","Data":"9157ef60d0943cd2af72cbd27b0db717de026d5a9d06ccf7b53800063e46296c"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.506400 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" event={"ID":"87a8de16-a4f9-4750-a6ea-3796295fa455","Type":"ContainerStarted","Data":"9776c1574b6af50abb7225f072fe7befea15fdbf9298ec2b1b6c03d0ae20339e"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.506436 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" event={"ID":"87a8de16-a4f9-4750-a6ea-3796295fa455","Type":"ContainerStarted","Data":"b731ac539fe4943d6475dc1790c8051214fe945fe1d4d23960021d38604f911a"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.517262 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" event={"ID":"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb","Type":"ContainerStarted","Data":"2723fbfbda28e70efb4fc6505b9788809b80099f0fd2796f8b82a9e721c1a32f"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.517294 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" event={"ID":"f3e8be91-d1a1-4107-bb4c-67a04cfd93cb","Type":"ContainerStarted","Data":"4ab33c30e9016d458fb7891f30debf8ef3d711af40c2cfcdc31140b5c4a5f5ad"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.526106 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" event={"ID":"e20fa67e-381d-49ff-82da-aff52e45d882","Type":"ContainerStarted","Data":"375fbc47fca1029c1f6fb9436f78b8e25ced99cc58f02211239c9ce3bf61d242"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.526408 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" event={"ID":"e20fa67e-381d-49ff-82da-aff52e45d882","Type":"ContainerStarted","Data":"7ef540a64dd79a45598e3ce3949be301df6ebe31c6d8b243f49c6c4ba5b827c6"} Nov 25 05:42:54 crc kubenswrapper[4708]: W1125 05:42:54.542573 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod392e81c8_e2ab_4bd2_97b1_ba0553e6b485.slice/crio-5a0dd07011754c2d19c8821a38fab139b22036764ae73a78b31154f0812b5d08 WatchSource:0}: Error finding container 5a0dd07011754c2d19c8821a38fab139b22036764ae73a78b31154f0812b5d08: Status 404 returned error can't find the container with id 5a0dd07011754c2d19c8821a38fab139b22036764ae73a78b31154f0812b5d08 Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.547851 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" event={"ID":"c35d3991-35e6-4cd7-a3c9-11513018d1fd","Type":"ContainerStarted","Data":"1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.555984 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" podStartSLOduration=98.555966624 podStartE2EDuration="1m38.555966624s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:54.544469747 +0000 UTC m=+115.953303132" watchObservedRunningTime="2025-11-25 05:42:54.555966624 +0000 UTC m=+115.964800011" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.558732 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vnvqz" event={"ID":"389881b2-1179-4817-8bd0-9628ddbdaf31","Type":"ContainerStarted","Data":"fb8ee6f4724c0a5fbc5e958e6e45dd5a30d472a0eb42c883a5f693df3fae242b"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.558773 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-vnvqz" event={"ID":"389881b2-1179-4817-8bd0-9628ddbdaf31","Type":"ContainerStarted","Data":"dfad71404dbc44f49e99e50dbf1f13004ddda657929f6d3da65a4e6f3972eec1"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.561509 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-t687t"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.566961 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.568017 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn"] Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.568827 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.068810096 +0000 UTC m=+116.477643482 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.570139 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" event={"ID":"7e29a355-332e-4f3b-a31a-20f5a0eaaaad","Type":"ContainerStarted","Data":"03678a480af7e8d6d92636911074d41935ca03a3af7ea145baa2fd2d00b6d02b"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.570171 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" event={"ID":"7e29a355-332e-4f3b-a31a-20f5a0eaaaad","Type":"ContainerStarted","Data":"235b28ab7e1433dd11fdc6d71ab5c74b67149778f343fa079abcaee7010ea31d"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.577475 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-6vkgb" podStartSLOduration=98.577462082 podStartE2EDuration="1m38.577462082s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:54.575499881 +0000 UTC m=+115.984333267" watchObservedRunningTime="2025-11-25 05:42:54.577462082 +0000 UTC m=+115.986295468" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.585850 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8"] Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.606948 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" event={"ID":"1634a636-2d11-464c-9306-0123e62c65d0","Type":"ContainerStarted","Data":"fe6c7e66398e1fb94f6e338a98b36f956be70ce9665e2a7c78a36bd252a42a42"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.618414 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" event={"ID":"fd80abac-33e8-498f-a346-7bc0e42886c1","Type":"ContainerStarted","Data":"9c7e8ca27663e379f7f5e50395bf021138cc5b1a06fa6c8b70aaf46ae62a895b"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.674183 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.678817 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" event={"ID":"0cfd3a04-2b25-4da4-a565-69b909c2d5e3","Type":"ContainerStarted","Data":"59a6e578acd9ca71697229095663dcc2df68d8f9bb7bfc5971b74e78ff15c1c5"} Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.679360 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.179092573 +0000 UTC m=+116.587925959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.687578 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" event={"ID":"4af5429b-294f-4028-8b55-1b55f11d7987","Type":"ContainerStarted","Data":"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.688039 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" event={"ID":"4af5429b-294f-4028-8b55-1b55f11d7987","Type":"ContainerStarted","Data":"84bc8b687987643fd6b723d22d0db3823288b8049c404bed01d348ae966f3d05"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.688511 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.709224 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" event={"ID":"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67","Type":"ContainerStarted","Data":"db549abe45700aa2b6f6851df287ada8be3f5640e17b4c514a5163097713dce6"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.709266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" event={"ID":"6fe7a4d8-8f9e-4f02-89ca-986cbcd76a67","Type":"ContainerStarted","Data":"afe48ba9d136edd5bd92f47095c42dd458af854b1f84e814cdfeb24ee83a5e7b"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.712144 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" event={"ID":"41359abe-343a-4909-b6bf-8bc71f24fc5e","Type":"ContainerStarted","Data":"f5874fcb5e7727f9415038c9d7347198a0bb939b4f553ef5cd2df7e556f3f3db"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.712190 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" event={"ID":"41359abe-343a-4909-b6bf-8bc71f24fc5e","Type":"ContainerStarted","Data":"4d5b9623359959fc73e1ed8baa019037ddf2e73af9944a97ab394237f940d933"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.713561 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r9f4b" event={"ID":"9a770195-b221-4ec8-bbbc-92573d42db74","Type":"ContainerStarted","Data":"ec72992276903871b7d48307f3a9b546022a7b303c3cad7263fe31f1463285f3"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.724080 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" event={"ID":"ac1c761e-3ccf-4c76-8e06-8e98390da65f","Type":"ContainerStarted","Data":"4c2d6ede09b82d162dec81853beb4cb4d99d71e060bce9d2456dd2b40afd3537"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.724121 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" event={"ID":"ac1c761e-3ccf-4c76-8e06-8e98390da65f","Type":"ContainerStarted","Data":"2ec8a3a0a6e4cd56527d21908e5765721f4794a3c8a7fcc5194782849e76fee1"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.724960 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.727167 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" event={"ID":"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21","Type":"ContainerStarted","Data":"9b3398d1167de77fabc7ab79fe430589bccf06e8a02f53e9a4bdcfe63ffa50f4"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.748958 4708 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-n578c container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.749235 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" podUID="ac1c761e-3ccf-4c76-8e06-8e98390da65f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.750787 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" event={"ID":"ec56d949-050a-459d-853e-5cbcf2347699","Type":"ContainerStarted","Data":"011b890c30a0841aa2bc3a6a5c69a7b3e85c2c1052dd90a43b9970c18b85cb7a"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.768719 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" event={"ID":"8a55393c-9185-4059-a3ed-f79eeb2f69bf","Type":"ContainerStarted","Data":"926f37141ed22eb56b93aeed643b06c1f3a1528e6bc9f381143c697cc4cb9f3d"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.768764 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" event={"ID":"8a55393c-9185-4059-a3ed-f79eeb2f69bf","Type":"ContainerStarted","Data":"0f06260806bb01933263ffd40dd600fbdf7fafe8d0221939d922240dc144c2fe"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.774962 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.775324 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.275281375 +0000 UTC m=+116.684114761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.803559 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" event={"ID":"e1dcf9d6-33d7-4482-b24f-990de83f2fa1","Type":"ContainerStarted","Data":"d927312fe6033e0c8db51ce0018cde051475efaed1bea0f61965739a5ae9bdcf"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.803592 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" event={"ID":"e1dcf9d6-33d7-4482-b24f-990de83f2fa1","Type":"ContainerStarted","Data":"0e9b98a6d2befdb72f5f4b8fcc82d36e47029b75c2dcbc44e8bb3c1aef2b2466"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.810266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" event={"ID":"062dced0-ecfd-49fa-bfbb-5ff51768816c","Type":"ContainerStarted","Data":"e9263139b032a253449be3bad64e29f740217bf951afac12168977d41f9c0deb"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.812454 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" event={"ID":"bcf83749-f980-4108-b8f5-a618548ccdbf","Type":"ContainerStarted","Data":"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.812496 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" event={"ID":"bcf83749-f980-4108-b8f5-a618548ccdbf","Type":"ContainerStarted","Data":"bb629ded7086b72a769a58c1ad26b5427b6ee9dd50ef9e653e1d0521b88b4bf2"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.812949 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.815933 4708 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-6gs2b container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.815973 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.816824 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" event={"ID":"7c9f0208-cc82-49bc-907c-7f9f81832fa7","Type":"ContainerStarted","Data":"30d0d384dd86fc1a319cd4c6e04ece3fef04db6c2e308ee15aa4ddb97e735c8a"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.826782 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" event={"ID":"c6b06363-6a06-4bb6-b112-383a4e95475e","Type":"ContainerStarted","Data":"b01a5e0a44010a01704c3397d0e8f2ae0a472b821b542f8f999d2ff270445992"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.863358 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r2zjk" event={"ID":"db4f79f6-d226-40c5-bf33-e0f545525833","Type":"ContainerStarted","Data":"85c85bd5925eb70b359eb2febdac4ac6433d97309f6128a2b06964622cd2d484"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.863417 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r2zjk" event={"ID":"db4f79f6-d226-40c5-bf33-e0f545525833","Type":"ContainerStarted","Data":"8ac057fa877ea6c3a634f69fdfd4cfad41952843d27127f4f85eb5b0cac3884c"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.874741 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" event={"ID":"4c24dca1-f55a-4893-a179-ad1ce776d138","Type":"ContainerStarted","Data":"a133054f38b9642566d8bd672b7a09caeaa0e38da9eb64e54c33e3f03bdf8666"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.879188 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.879830 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.379811541 +0000 UTC m=+116.788644927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.886951 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" event={"ID":"e41814f6-cdb7-4460-bc75-618b5fff39df","Type":"ContainerStarted","Data":"d2555b6a50fcb5f126c7e809665865f1e8f323df49f743fed52eb1fce4cb1950"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.899300 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-snrvr" podStartSLOduration=98.899274535 podStartE2EDuration="1m38.899274535s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:54.895813013 +0000 UTC m=+116.304646399" watchObservedRunningTime="2025-11-25 05:42:54.899274535 +0000 UTC m=+116.308107920" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.907376 4708 patch_prober.go:28] interesting pod/downloads-7954f5f757-qtrmv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.907622 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qtrmv" podUID="aa6199ad-a3fd-4547-83ca-06944728979d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.909346 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.909367 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qtrmv" event={"ID":"aa6199ad-a3fd-4547-83ca-06944728979d","Type":"ContainerStarted","Data":"467ec8129f7d28d1abb8e6df94e78a8122a763dc12d17c937cc8f5472c0e98db"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.913817 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" event={"ID":"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c","Type":"ContainerStarted","Data":"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc"} Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.913882 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.916846 4708 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-2kwv6 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" start-of-body= Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.916879 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.29:6443/healthz\": dial tcp 10.217.0.29:6443: connect: connection refused" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.923546 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-z6ctz" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.925413 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:42:54 crc kubenswrapper[4708]: I1125 05:42:54.981479 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:54 crc kubenswrapper[4708]: E1125 05:42:54.984217 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.484200359 +0000 UTC m=+116.893033745 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.013042 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" podStartSLOduration=98.013029754 podStartE2EDuration="1m38.013029754s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.012118029 +0000 UTC m=+116.420951416" watchObservedRunningTime="2025-11-25 05:42:55.013029754 +0000 UTC m=+116.421863140" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.084901 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.085884 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.086231 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.586220914 +0000 UTC m=+116.995054300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.107793 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" podStartSLOduration=98.107780732 podStartE2EDuration="1m38.107780732s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.10472837 +0000 UTC m=+116.513561756" watchObservedRunningTime="2025-11-25 05:42:55.107780732 +0000 UTC m=+116.516614118" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.140926 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" podStartSLOduration=98.140911178 podStartE2EDuration="1m38.140911178s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.135381606 +0000 UTC m=+116.544214992" watchObservedRunningTime="2025-11-25 05:42:55.140911178 +0000 UTC m=+116.549744564" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.171327 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" podStartSLOduration=99.171315776 podStartE2EDuration="1m39.171315776s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.169285096 +0000 UTC m=+116.578118482" watchObservedRunningTime="2025-11-25 05:42:55.171315776 +0000 UTC m=+116.580149183" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.189597 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.190115 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.690101506 +0000 UTC m=+117.098934883 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.275376 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-642cw" podStartSLOduration=98.275360027 podStartE2EDuration="1m38.275360027s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.273694955 +0000 UTC m=+116.682528341" watchObservedRunningTime="2025-11-25 05:42:55.275360027 +0000 UTC m=+116.684193413" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.291716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.292137 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.792124444 +0000 UTC m=+117.200957831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.344410 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" podStartSLOduration=99.344386332 podStartE2EDuration="1m39.344386332s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.343391831 +0000 UTC m=+116.752225206" watchObservedRunningTime="2025-11-25 05:42:55.344386332 +0000 UTC m=+116.753219717" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.345168 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r2zjk" podStartSLOduration=99.345159236 podStartE2EDuration="1m39.345159236s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.296647855 +0000 UTC m=+116.705481241" watchObservedRunningTime="2025-11-25 05:42:55.345159236 +0000 UTC m=+116.753992622" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.371574 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-6db8s" podStartSLOduration=99.371554521 podStartE2EDuration="1m39.371554521s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.369500607 +0000 UTC m=+116.778333993" watchObservedRunningTime="2025-11-25 05:42:55.371554521 +0000 UTC m=+116.780387907" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.393201 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.393532 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.893501458 +0000 UTC m=+117.302334844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.393803 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.394137 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:55.894126003 +0000 UTC m=+117.302959389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.434190 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rzbf4" podStartSLOduration=99.434166589 podStartE2EDuration="1m39.434166589s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.432981499 +0000 UTC m=+116.841814885" watchObservedRunningTime="2025-11-25 05:42:55.434166589 +0000 UTC m=+116.842999974" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.452807 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.454292 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-49trj" podStartSLOduration=98.454273143 podStartE2EDuration="1m38.454273143s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.452765998 +0000 UTC m=+116.861599384" watchObservedRunningTime="2025-11-25 05:42:55.454273143 +0000 UTC m=+116.863106519" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.463677 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:42:55 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:42:55 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:42:55 crc kubenswrapper[4708]: healthz check failed Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.463714 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.500881 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.501338 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.001322193 +0000 UTC m=+117.410155579 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.537337 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qtrmv" podStartSLOduration=99.537315124 podStartE2EDuration="1m39.537315124s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.537079831 +0000 UTC m=+116.945913217" watchObservedRunningTime="2025-11-25 05:42:55.537315124 +0000 UTC m=+116.946148511" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.537945 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-5tpvl" podStartSLOduration=99.537938597 podStartE2EDuration="1m39.537938597s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.498317621 +0000 UTC m=+116.907151007" watchObservedRunningTime="2025-11-25 05:42:55.537938597 +0000 UTC m=+116.946771983" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.580927 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-4wdjr" podStartSLOduration=98.580902541 podStartE2EDuration="1m38.580902541s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.580021014 +0000 UTC m=+116.988854399" watchObservedRunningTime="2025-11-25 05:42:55.580902541 +0000 UTC m=+116.989735928" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.611328 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.611981 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.111970097 +0000 UTC m=+117.520803483 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.649286 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-vnvqz" podStartSLOduration=5.649266229 podStartE2EDuration="5.649266229s" podCreationTimestamp="2025-11-25 05:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.610639094 +0000 UTC m=+117.019472479" watchObservedRunningTime="2025-11-25 05:42:55.649266229 +0000 UTC m=+117.058099616" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.696485 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-97g7p" podStartSLOduration=99.69646417 podStartE2EDuration="1m39.69646417s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.651313502 +0000 UTC m=+117.060146888" watchObservedRunningTime="2025-11-25 05:42:55.69646417 +0000 UTC m=+117.105297556" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.713047 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.713483 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.213469651 +0000 UTC m=+117.622303037 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.735448 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kgmw8" podStartSLOduration=98.735429963 podStartE2EDuration="1m38.735429963s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.734379026 +0000 UTC m=+117.143212412" watchObservedRunningTime="2025-11-25 05:42:55.735429963 +0000 UTC m=+117.144263349" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.815833 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.815803 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-n58zc" podStartSLOduration=99.815783707 podStartE2EDuration="1m39.815783707s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.779551717 +0000 UTC m=+117.188385103" watchObservedRunningTime="2025-11-25 05:42:55.815783707 +0000 UTC m=+117.224617093" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.816327 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.316310508 +0000 UTC m=+117.725143914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.916965 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.917144 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.417113041 +0000 UTC m=+117.825946426 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.917317 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:55 crc kubenswrapper[4708]: E1125 05:42:55.917651 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.417636205 +0000 UTC m=+117.826469591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.921932 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" event={"ID":"37759660-fe73-4292-b49d-e7572f3838ee","Type":"ContainerStarted","Data":"5fb9e7a94e52fef8e593fa7575d53bc75d4ef84d91c4bc9c10c145642665130c"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.921984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" event={"ID":"37759660-fe73-4292-b49d-e7572f3838ee","Type":"ContainerStarted","Data":"a52138b69ae49d8a42960fe52fdb91af2c7e8f2bdbbb2ccb9e18f9ea4500c84d"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.925606 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" event={"ID":"9edd4399-bff5-4125-be89-3519a2d94ab4","Type":"ContainerStarted","Data":"fc3e2b76c9071ccb066da0da2cccee57d6775fdf6ff24c989c563c3258465f54"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.925640 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" event={"ID":"9edd4399-bff5-4125-be89-3519a2d94ab4","Type":"ContainerStarted","Data":"854d53176ba3de596ab90583b403d97702c47569a6017365fd342da527eb9561"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.926447 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.927909 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" event={"ID":"46c16b8d-15a9-4d6c-a726-f7d845043a13","Type":"ContainerStarted","Data":"046955d64664ce2940396f15756b2b1fbaa242774bdab91fdd9b22765271885f"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.930770 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" event={"ID":"46c16b8d-15a9-4d6c-a726-f7d845043a13","Type":"ContainerStarted","Data":"ba3848c19f1278a1b2174f02c3642095d9e1d62e831c8bd05a06d69b7d8e9d69"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.931640 4708 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vjzpn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" start-of-body= Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.931685 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" podUID="9edd4399-bff5-4125-be89-3519a2d94ab4" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.40:5443/healthz\": dial tcp 10.217.0.40:5443: connect: connection refused" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.944183 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" event={"ID":"e41814f6-cdb7-4460-bc75-618b5fff39df","Type":"ContainerStarted","Data":"2f53e9e8bd13294d5f93d431c1442ee0a658a25f1d17aa0b083040a595ccee11"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.944224 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" event={"ID":"e41814f6-cdb7-4460-bc75-618b5fff39df","Type":"ContainerStarted","Data":"0c3f1071b907fc3ea7759b8b943395811e94b231b442871b68bbae9f84214936"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.947345 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" event={"ID":"fd80abac-33e8-498f-a346-7bc0e42886c1","Type":"ContainerStarted","Data":"9d3f0ead1d6bfe986a514029a8ad53499575bdbfacec98b5859552554c47fdcc"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.953890 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" event={"ID":"9ac43907-bde5-455d-81a1-71264ac17d51","Type":"ContainerStarted","Data":"1b0af366f7b790ec3f8591c326d55211044974cdeb43eb6105a9f7c8e85f7540"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.953919 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" event={"ID":"9ac43907-bde5-455d-81a1-71264ac17d51","Type":"ContainerStarted","Data":"391e4d285a8ff6dcef9cfbab2f45b0819565aedd6b3186922e87c6b53e417ddf"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.959402 4708 generic.go:334] "Generic (PLEG): container finished" podID="c35d3991-35e6-4cd7-a3c9-11513018d1fd" containerID="35d229d1a04c1aa30ecf7da1cf037f3f78d5e52658ddc23b31ba457a7eeadcad" exitCode=0 Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.959574 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" event={"ID":"c35d3991-35e6-4cd7-a3c9-11513018d1fd","Type":"ContainerDied","Data":"35d229d1a04c1aa30ecf7da1cf037f3f78d5e52658ddc23b31ba457a7eeadcad"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.959789 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-mhxgn" podStartSLOduration=99.959772543 podStartE2EDuration="1m39.959772543s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.958815182 +0000 UTC m=+117.367648568" watchObservedRunningTime="2025-11-25 05:42:55.959772543 +0000 UTC m=+117.368605929" Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.964881 4708 generic.go:334] "Generic (PLEG): container finished" podID="0cfd3a04-2b25-4da4-a565-69b909c2d5e3" containerID="1f65f0b41bd27ede3ca9a319a94f265e1f53f1dadb40b1f3735c9784a9d1d28c" exitCode=0 Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.964942 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" event={"ID":"0cfd3a04-2b25-4da4-a565-69b909c2d5e3","Type":"ContainerDied","Data":"1f65f0b41bd27ede3ca9a319a94f265e1f53f1dadb40b1f3735c9784a9d1d28c"} Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.975562 4708 generic.go:334] "Generic (PLEG): container finished" podID="c6b06363-6a06-4bb6-b112-383a4e95475e" containerID="a2f50c126cff6acf3020d94cf386830551cc36ac06306d054a8411b3f69a776b" exitCode=0 Nov 25 05:42:55 crc kubenswrapper[4708]: I1125 05:42:55.975630 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" event={"ID":"c6b06363-6a06-4bb6-b112-383a4e95475e","Type":"ContainerDied","Data":"a2f50c126cff6acf3020d94cf386830551cc36ac06306d054a8411b3f69a776b"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.011228 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" event={"ID":"847c28aa-0945-492d-941c-b29304b7768b","Type":"ContainerStarted","Data":"d4d6785529bda686760a583c2e873281dc3af2a84e4d33fae65c53c26174ea0f"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.011272 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" event={"ID":"847c28aa-0945-492d-941c-b29304b7768b","Type":"ContainerStarted","Data":"afc1e948c86ed3b1d12b0e48df7e8d819747b61c86a30448f9da2e647d3274f2"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.011284 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" event={"ID":"847c28aa-0945-492d-941c-b29304b7768b","Type":"ContainerStarted","Data":"d468cd2da1db8822c7411ea8b4bbafb5477697e1fc86db57437869ee9c2a8c27"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.011800 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.014895 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" event={"ID":"4c24dca1-f55a-4893-a179-ad1ce776d138","Type":"ContainerStarted","Data":"b7259cf7ea0d0c224e6d126375e0e8121ecfd61dd991f666d0e6526c276790ec"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.021346 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" podStartSLOduration=99.021331619 podStartE2EDuration="1m39.021331619s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:55.988221802 +0000 UTC m=+117.397055188" watchObservedRunningTime="2025-11-25 05:42:56.021331619 +0000 UTC m=+117.430165005" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.028954 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-t687t" event={"ID":"10e41c07-da3d-4999-8919-5a7e6102a22d","Type":"ContainerStarted","Data":"919076e7b8d6f74339fd01d877712491e8d5acbd03e1aad418bcbd70ce086047"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.028986 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-t687t" event={"ID":"10e41c07-da3d-4999-8919-5a7e6102a22d","Type":"ContainerStarted","Data":"6070a47ee2ed683d5396418d1599aba83e25fc8dce4da4e33ba05c6c4a0fdcb1"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.029098 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.030208 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.530183211 +0000 UTC m=+117.939016598 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.050214 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" event={"ID":"7c9f0208-cc82-49bc-907c-7f9f81832fa7","Type":"ContainerStarted","Data":"e02d82bfc636498d7a516b6248171d318d3b1b2f23146eaa271a0dd3c3dbd137"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.051013 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-f98lb" podStartSLOduration=99.050987819 podStartE2EDuration="1m39.050987819s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.050873224 +0000 UTC m=+117.459706610" watchObservedRunningTime="2025-11-25 05:42:56.050987819 +0000 UTC m=+117.459821205" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.053012 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-kcnvk" podStartSLOduration=99.053007169 podStartE2EDuration="1m39.053007169s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.022155148 +0000 UTC m=+117.430988534" watchObservedRunningTime="2025-11-25 05:42:56.053007169 +0000 UTC m=+117.461840554" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.068730 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" event={"ID":"1634a636-2d11-464c-9306-0123e62c65d0","Type":"ContainerStarted","Data":"6304d9e8b24fc7237085fb024ffa0392ebef9b8d8eefbd4385bb760cc7000787"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.081419 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-sskm9" podStartSLOduration=100.081403148 podStartE2EDuration="1m40.081403148s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.078825879 +0000 UTC m=+117.487659265" watchObservedRunningTime="2025-11-25 05:42:56.081403148 +0000 UTC m=+117.490236533" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.084135 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" event={"ID":"41359abe-343a-4909-b6bf-8bc71f24fc5e","Type":"ContainerStarted","Data":"654ee7a6574a14c57904239fa891389983d60106d4d57ad9c32ea2d21e33351d"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.086007 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" event={"ID":"f9d144cb-ff11-40a4-a30a-d4feb8c3d58c","Type":"ContainerStarted","Data":"763760b922515962e4ef2b9a08900371dd116899cd8ceab1ecf7fa475a2d3e7b"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.087374 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r9f4b" event={"ID":"9a770195-b221-4ec8-bbbc-92573d42db74","Type":"ContainerStarted","Data":"67466c399380b3ec81ad7951057fcca8185cf01cd48656fef85bd6d5f9a521e7"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.109556 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" podStartSLOduration=100.109540641 podStartE2EDuration="1m40.109540641s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.106083257 +0000 UTC m=+117.514916643" watchObservedRunningTime="2025-11-25 05:42:56.109540641 +0000 UTC m=+117.518374027" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.113663 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" event={"ID":"392e81c8-e2ab-4bd2-97b1-ba0553e6b485","Type":"ContainerStarted","Data":"0f45de0ac27ce0a6900c7eb880ceaf0bf92353ed6800b94dc00f790beafafcf4"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.113698 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" event={"ID":"392e81c8-e2ab-4bd2-97b1-ba0553e6b485","Type":"ContainerStarted","Data":"5a0dd07011754c2d19c8821a38fab139b22036764ae73a78b31154f0812b5d08"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.122852 4708 generic.go:334] "Generic (PLEG): container finished" podID="c743ab1a-6a44-49e0-b9ec-ccf8b8213a21" containerID="73db869f69b7566f2a636b5b2e221cadc9507c07c7674f21e03e1afda29f56b6" exitCode=0 Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.122909 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" event={"ID":"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21","Type":"ContainerDied","Data":"73db869f69b7566f2a636b5b2e221cadc9507c07c7674f21e03e1afda29f56b6"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.126813 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2fgg4" podStartSLOduration=100.126804047 podStartE2EDuration="1m40.126804047s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.126214608 +0000 UTC m=+117.535047994" watchObservedRunningTime="2025-11-25 05:42:56.126804047 +0000 UTC m=+117.535637434" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.131202 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.133080 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.633067369 +0000 UTC m=+118.041900756 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.171888 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" event={"ID":"76500481-c48a-4cf4-993e-ba1f81dc4327","Type":"ContainerStarted","Data":"eb01224924c846261189c5fbfd124903aaa142c67a27de13070d4426e1cc47e3"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.171935 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" event={"ID":"76500481-c48a-4cf4-993e-ba1f81dc4327","Type":"ContainerStarted","Data":"fa887a2b72f1f7df7a3f32f9266ddfeea537e4410ccbb73046d6ea9ab729b5f1"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.172588 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.189950 4708 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-schhr container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" start-of-body= Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.190006 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" podUID="76500481-c48a-4cf4-993e-ba1f81dc4327" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.34:8443/healthz\": dial tcp 10.217.0.34:8443: connect: connection refused" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.200128 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkmtb" event={"ID":"c3776f93-7ce9-4056-86c8-7a2d95a02650","Type":"ContainerStarted","Data":"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45"} Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.200390 4708 patch_prober.go:28] interesting pod/downloads-7954f5f757-qtrmv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" start-of-body= Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.200427 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qtrmv" podUID="aa6199ad-a3fd-4547-83ca-06944728979d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.7:8080/\": dial tcp 10.217.0.7:8080: connect: connection refused" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.221810 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.221859 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.226839 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-n578c" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.233009 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.233204 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.733178192 +0000 UTC m=+118.142011577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.233572 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.235263 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.73525665 +0000 UTC m=+118.144090037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.338232 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.338384 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.838362146 +0000 UTC m=+118.247195532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.338937 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.342665 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.842639261 +0000 UTC m=+118.251472648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.385453 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" podStartSLOduration=99.385433888 podStartE2EDuration="1m39.385433888s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.291744848 +0000 UTC m=+117.700578234" watchObservedRunningTime="2025-11-25 05:42:56.385433888 +0000 UTC m=+117.794267275" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.441855 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.442283 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:56.942268788 +0000 UTC m=+118.351102175 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.456190 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:42:56 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:42:56 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:42:56 crc kubenswrapper[4708]: healthz check failed Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.456248 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.514954 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-t687t" podStartSLOduration=6.514934159 podStartE2EDuration="6.514934159s" podCreationTimestamp="2025-11-25 05:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.388664296 +0000 UTC m=+117.797497672" watchObservedRunningTime="2025-11-25 05:42:56.514934159 +0000 UTC m=+117.923767715" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.520639 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-6gm6k" podStartSLOduration=99.520623632 podStartE2EDuration="1m39.520623632s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.51389893 +0000 UTC m=+117.922732316" watchObservedRunningTime="2025-11-25 05:42:56.520623632 +0000 UTC m=+117.929457018" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.543227 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.543594 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.043580739 +0000 UTC m=+118.452414126 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.564546 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-z4frk" podStartSLOduration=100.564532233 podStartE2EDuration="1m40.564532233s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.564242719 +0000 UTC m=+117.973076104" watchObservedRunningTime="2025-11-25 05:42:56.564532233 +0000 UTC m=+117.973365620" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.606565 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q294h" podStartSLOduration=100.606548826 podStartE2EDuration="1m40.606548826s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.605812231 +0000 UTC m=+118.014645617" watchObservedRunningTime="2025-11-25 05:42:56.606548826 +0000 UTC m=+118.015382212" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.633151 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-l52tr" podStartSLOduration=99.633137074 podStartE2EDuration="1m39.633137074s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.631430034 +0000 UTC m=+118.040263419" watchObservedRunningTime="2025-11-25 05:42:56.633137074 +0000 UTC m=+118.041970460" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.644340 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.644634 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.144596413 +0000 UTC m=+118.553429799 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.644850 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.645226 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.14521198 +0000 UTC m=+118.554045366 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.747909 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.748282 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.248267922 +0000 UTC m=+118.657101309 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.770280 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zkmtb" podStartSLOduration=100.77025765 podStartE2EDuration="1m40.77025765s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.739383809 +0000 UTC m=+118.148217195" watchObservedRunningTime="2025-11-25 05:42:56.77025765 +0000 UTC m=+118.179091036" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.778775 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" podStartSLOduration=99.778756448 podStartE2EDuration="1m39.778756448s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:56.778462555 +0000 UTC m=+118.187295931" watchObservedRunningTime="2025-11-25 05:42:56.778756448 +0000 UTC m=+118.187589834" Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.850215 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.850565 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.350550238 +0000 UTC m=+118.759383625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:56 crc kubenswrapper[4708]: I1125 05:42:56.951998 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:56 crc kubenswrapper[4708]: E1125 05:42:56.952470 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.452450446 +0000 UTC m=+118.861283832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.053898 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.054335 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.554317441 +0000 UTC m=+118.963150827 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.156013 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.156155 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.65612828 +0000 UTC m=+119.064961667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.156306 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.156776 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.656755801 +0000 UTC m=+119.065589187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.205255 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" event={"ID":"392e81c8-e2ab-4bd2-97b1-ba0553e6b485","Type":"ContainerStarted","Data":"915fcd99c21febfd90db9ec1b9b04697f824cbeab4edd297029ee93334edf455"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.205302 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" event={"ID":"392e81c8-e2ab-4bd2-97b1-ba0553e6b485","Type":"ContainerStarted","Data":"f2ee6741d955bdf4d9b4e28df1f91d211270f1672a2d0c7e7b107601f3fd3236"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.206956 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-nlcx6" event={"ID":"9ac43907-bde5-455d-81a1-71264ac17d51","Type":"ContainerStarted","Data":"1687970d77bebbc7c19ad51a420eda68b7648d128b8c97cc287e112b6c9422d7"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.209169 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" event={"ID":"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21","Type":"ContainerStarted","Data":"b4b3b7fdd7b9dcb9bd9cc4393278369e07723a569ed9f974490d6c6e7597962a"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.209204 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" event={"ID":"c743ab1a-6a44-49e0-b9ec-ccf8b8213a21","Type":"ContainerStarted","Data":"6a3ab3d86bfca2c63fbcd55f94df85774bb50d4247590db75490022f30855a5b"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.211197 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" event={"ID":"c6b06363-6a06-4bb6-b112-383a4e95475e","Type":"ContainerStarted","Data":"4e26dd288696c5ace316a33e9f393c91f413adf3f071b115bfd5d764afa15cba"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.211562 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.212816 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" event={"ID":"0cfd3a04-2b25-4da4-a565-69b909c2d5e3","Type":"ContainerStarted","Data":"9e00a0d86ce972fd862eb677a75baea30178b7ade441a8995fcda3bc604e2645"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.215455 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-r9f4b" event={"ID":"9a770195-b221-4ec8-bbbc-92573d42db74","Type":"ContainerStarted","Data":"cf1721439d444af50e5506b1708a5a49f02c5a350cb8ec93845ed15959133dd0"} Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.223665 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vjzpn" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.236384 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-schhr" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.257996 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.258327 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.758311169 +0000 UTC m=+119.167144545 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.293573 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" podStartSLOduration=101.293556584 podStartE2EDuration="1m41.293556584s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:57.291786093 +0000 UTC m=+118.700619479" watchObservedRunningTime="2025-11-25 05:42:57.293556584 +0000 UTC m=+118.702389970" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.372094 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.381323 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.881309297 +0000 UTC m=+119.290142683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.450775 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-r9f4b" podStartSLOduration=7.450756072 podStartE2EDuration="7.450756072s" podCreationTimestamp="2025-11-25 05:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:57.411038865 +0000 UTC m=+118.819872240" watchObservedRunningTime="2025-11-25 05:42:57.450756072 +0000 UTC m=+118.859589458" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.451373 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" podStartSLOduration=100.451369987 podStartE2EDuration="1m40.451369987s" podCreationTimestamp="2025-11-25 05:41:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:57.450307678 +0000 UTC m=+118.859141064" watchObservedRunningTime="2025-11-25 05:42:57.451369987 +0000 UTC m=+118.860203373" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.468304 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:42:57 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:42:57 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:42:57 crc kubenswrapper[4708]: healthz check failed Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.468361 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.480260 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" podStartSLOduration=101.480242884 podStartE2EDuration="1m41.480242884s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:57.477896329 +0000 UTC m=+118.886729715" watchObservedRunningTime="2025-11-25 05:42:57.480242884 +0000 UTC m=+118.889076259" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.480347 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.980329066 +0000 UTC m=+119.389162442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.480281 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.482050 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.482442 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:57.982427042 +0000 UTC m=+119.391260428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.583940 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.584459 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.084442237 +0000 UTC m=+119.493275623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.685089 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.688362 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.688782 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.188769269 +0000 UTC m=+119.597602655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.789882 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.790071 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume\") pod \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.790131 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.290079256 +0000 UTC m=+119.698912643 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.790209 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9gg5\" (UniqueName: \"kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5\") pod \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.790279 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume\") pod \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\" (UID: \"c35d3991-35e6-4cd7-a3c9-11513018d1fd\") " Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.790928 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.790974 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume" (OuterVolumeSpecName: "config-volume") pod "c35d3991-35e6-4cd7-a3c9-11513018d1fd" (UID: "c35d3991-35e6-4cd7-a3c9-11513018d1fd"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.791129 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c35d3991-35e6-4cd7-a3c9-11513018d1fd-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.791216 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.291200064 +0000 UTC m=+119.700033450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.809644 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5" (OuterVolumeSpecName: "kube-api-access-n9gg5") pod "c35d3991-35e6-4cd7-a3c9-11513018d1fd" (UID: "c35d3991-35e6-4cd7-a3c9-11513018d1fd"). InnerVolumeSpecName "kube-api-access-n9gg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.810018 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c35d3991-35e6-4cd7-a3c9-11513018d1fd" (UID: "c35d3991-35e6-4cd7-a3c9-11513018d1fd"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.892307 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.892512 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.392470628 +0000 UTC m=+119.801304014 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.893036 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.893426 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.39340724 +0000 UTC m=+119.802240626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.893786 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c35d3991-35e6-4cd7-a3c9-11513018d1fd-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.893861 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9gg5\" (UniqueName: \"kubernetes.io/projected/c35d3991-35e6-4cd7-a3c9-11513018d1fd-kube-api-access-n9gg5\") on node \"crc\" DevicePath \"\"" Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.994593 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.994874 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.494844807 +0000 UTC m=+119.903678192 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:57 crc kubenswrapper[4708]: I1125 05:42:57.995143 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:57 crc kubenswrapper[4708]: E1125 05:42:57.995570 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.495560062 +0000 UTC m=+119.904393449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.030311 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.030355 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.095963 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.096129 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.596109288 +0000 UTC m=+120.004942675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.096309 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.096674 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.596666697 +0000 UTC m=+120.005500083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.126355 4708 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.159863 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.160051 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.167575 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.197264 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.197380 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.697358671 +0000 UTC m=+120.106192058 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.197790 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.198084 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.69807585 +0000 UTC m=+120.106909236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.221692 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" event={"ID":"392e81c8-e2ab-4bd2-97b1-ba0553e6b485","Type":"ContainerStarted","Data":"3b64de245a6fa1171ad65ed40ea62395ba299b3e2c59c44605b2afac157e7040"} Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.222791 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.222893 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r" event={"ID":"c35d3991-35e6-4cd7-a3c9-11513018d1fd","Type":"ContainerDied","Data":"1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd"} Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.222985 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e9a7633bce16e9f08f652cbbcb86916bcda010a1d844d1caefe53e45d58ddfd" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.224557 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-r9f4b" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.232235 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-kgp2r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.263577 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-nkt5s" podStartSLOduration=8.263560201 podStartE2EDuration="8.263560201s" podCreationTimestamp="2025-11-25 05:42:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:42:58.249013125 +0000 UTC m=+119.657846511" watchObservedRunningTime="2025-11-25 05:42:58.263560201 +0000 UTC m=+119.672393587" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.299209 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.299384 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.799357935 +0000 UTC m=+120.208191321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.299709 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.300152 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.800135788 +0000 UTC m=+120.208969175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.313124 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.313334 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c35d3991-35e6-4cd7-a3c9-11513018d1fd" containerName="collect-profiles" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.313354 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c35d3991-35e6-4cd7-a3c9-11513018d1fd" containerName="collect-profiles" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.313457 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c35d3991-35e6-4cd7-a3c9-11513018d1fd" containerName="collect-profiles" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.314144 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.316016 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.354439 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.400447 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.400616 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.900586549 +0000 UTC m=+120.309419945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.401383 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.401844 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:58.901826552 +0000 UTC m=+120.310659938 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.461095 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:42:58 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:42:58 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:42:58 crc kubenswrapper[4708]: healthz check failed Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.461366 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.502560 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.502682 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.002658139 +0000 UTC m=+120.411491525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.502921 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.503047 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.503078 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v95j7\" (UniqueName: \"kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.503113 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.503440 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.003425702 +0000 UTC m=+120.412259089 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.508797 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.510345 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.513804 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.519240 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.604558 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.604753 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.104720421 +0000 UTC m=+120.513553807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.604945 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.604976 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605001 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605139 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605175 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v95j7\" (UniqueName: \"kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605214 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605264 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n5gk\" (UniqueName: \"kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.605474 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.606474 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.615933 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.115919961 +0000 UTC m=+120.524753347 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.636945 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v95j7\" (UniqueName: \"kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7\") pod \"certified-operators-m5q65\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.707132 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.707383 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n5gk\" (UniqueName: \"kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.707427 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.707446 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.707984 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.708018 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.207998371 +0000 UTC m=+120.616831756 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.708207 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.709986 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.711090 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.727609 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.740323 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n5gk\" (UniqueName: \"kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk\") pod \"community-operators-fz4gp\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.809024 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.809081 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.809109 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.809176 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7r5pb\" (UniqueName: \"kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.809725 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.309704162 +0000 UTC m=+120.718537548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.824117 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.909700 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.910152 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.910380 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7r5pb\" (UniqueName: \"kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.910502 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.910562 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: E1125 05:42:58.910797 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.41078046 +0000 UTC m=+120.819613846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.911344 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.911835 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.914790 4708 patch_prober.go:28] interesting pod/apiserver-76f77b778f-qhzn8 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]log ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]etcd ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/max-in-flight-filter ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/image.openshift.io-apiserver-caches ok Nov 25 05:42:58 crc kubenswrapper[4708]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Nov 25 05:42:58 crc kubenswrapper[4708]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/project.openshift.io-projectcache ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/openshift.io-startinformers ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/openshift.io-restmapperupdater ok Nov 25 05:42:58 crc kubenswrapper[4708]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 05:42:58 crc kubenswrapper[4708]: livez check failed Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.914846 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" podUID="c743ab1a-6a44-49e0-b9ec-ccf8b8213a21" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.917944 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.926492 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.944071 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.946950 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7r5pb\" (UniqueName: \"kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb\") pod \"certified-operators-sjf6r\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:58 crc kubenswrapper[4708]: I1125 05:42:58.947009 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.005926 4708 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T05:42:58.126443334Z","Handler":null,"Name":""} Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.015645 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.015704 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcwzq\" (UniqueName: \"kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.015762 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.015799 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: E1125 05:42:59.016179 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 05:42:59.516165553 +0000 UTC m=+120.924998929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hmtsm" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.024667 4708 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.024708 4708 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.024845 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.029448 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:42:59 crc kubenswrapper[4708]: W1125 05:42:59.042791 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1cb9071_9924_4611_b46b_6783b59eb4cd.slice/crio-0b9b70c34341ab06f17f7243ceb62198defeacafdec869f255146e703065320f WatchSource:0}: Error finding container 0b9b70c34341ab06f17f7243ceb62198defeacafdec869f255146e703065320f: Status 404 returned error can't find the container with id 0b9b70c34341ab06f17f7243ceb62198defeacafdec869f255146e703065320f Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.103866 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:42:59 crc kubenswrapper[4708]: W1125 05:42:59.109320 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03bf86cb_ba81_4f29_92b8_8a3456e5d822.slice/crio-4d82e5ead400db115aca12c59894e8df50079e9d4f225877ced4aabb39bbbdc2 WatchSource:0}: Error finding container 4d82e5ead400db115aca12c59894e8df50079e9d4f225877ced4aabb39bbbdc2: Status 404 returned error can't find the container with id 4d82e5ead400db115aca12c59894e8df50079e9d4f225877ced4aabb39bbbdc2 Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.117268 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.117690 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcwzq\" (UniqueName: \"kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.117787 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.117878 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.118258 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.118701 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.129840 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.133585 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcwzq\" (UniqueName: \"kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq\") pod \"community-operators-xfh2t\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.197700 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.219916 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.224232 4708 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.224264 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.236487 4708 generic.go:334] "Generic (PLEG): container finished" podID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerID="5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241" exitCode=0 Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.236571 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerDied","Data":"5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241"} Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.236604 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerStarted","Data":"0b9b70c34341ab06f17f7243ceb62198defeacafdec869f255146e703065320f"} Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.238050 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.239703 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerStarted","Data":"4d82e5ead400db115aca12c59894e8df50079e9d4f225877ced4aabb39bbbdc2"} Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.240859 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hmtsm\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.245805 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-gqbkc" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.256169 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.309086 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.318503 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.430334 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.455197 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:42:59 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:42:59 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:42:59 crc kubenswrapper[4708]: healthz check failed Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.455287 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:42:59 crc kubenswrapper[4708]: I1125 05:42:59.481440 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.252974 4708 generic.go:334] "Generic (PLEG): container finished" podID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerID="1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95" exitCode=0 Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.253033 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerDied","Data":"1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.261951 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ca0d416-1d1f-49b7-869c-720144da0487" containerID="4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7" exitCode=0 Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.262063 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerDied","Data":"4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.262179 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerStarted","Data":"fc8c7c55b4dd7147d28e23b59994d6998fa87b11fdf951dd9e08ea22042dba75"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.264748 4708 generic.go:334] "Generic (PLEG): container finished" podID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerID="80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68" exitCode=0 Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.264830 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerDied","Data":"80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.264886 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerStarted","Data":"e8566f742cc1c83f9f444b7290bae8d5efb62a4386d998545d130c52c2b35cb5"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.269733 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" event={"ID":"e27e8c69-3b12-4193-9fc3-26d339707d95","Type":"ContainerStarted","Data":"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.269841 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" event={"ID":"e27e8c69-3b12-4193-9fc3-26d339707d95","Type":"ContainerStarted","Data":"a76c9f9d3a4490900ba7f87279bdb08b99883fa71294df5f635bc5fe77cfd076"} Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.305170 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.306148 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.307598 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.322380 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.323991 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" podStartSLOduration=104.32396978 podStartE2EDuration="1m44.32396978s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:43:00.31178559 +0000 UTC m=+121.720618976" watchObservedRunningTime="2025-11-25 05:43:00.32396978 +0000 UTC m=+121.732803166" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.435343 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.436024 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtk2r\" (UniqueName: \"kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.436188 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.461696 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:43:00 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:43:00 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:43:00 crc kubenswrapper[4708]: healthz check failed Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.461748 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.465499 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.468785 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.470331 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.472259 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.472535 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.538135 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtk2r\" (UniqueName: \"kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.538270 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.538332 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.538636 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.539610 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.554205 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtk2r\" (UniqueName: \"kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r\") pod \"redhat-marketplace-x6m7c\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.623815 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.639021 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.639170 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.709254 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.716487 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.727115 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.741511 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.741565 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.741612 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.756609 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.787192 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.802681 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.842720 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.842877 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.842957 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xxzr\" (UniqueName: \"kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.902126 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.944704 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.944787 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.944832 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xxzr\" (UniqueName: \"kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.945451 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.945714 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.967553 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xxzr\" (UniqueName: \"kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr\") pod \"redhat-marketplace-h47dj\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:00 crc kubenswrapper[4708]: I1125 05:43:00.976768 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 05:43:00 crc kubenswrapper[4708]: W1125 05:43:00.994184 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod078cd5e4_5781_4dc8_964e_6b1571ac85f9.slice/crio-25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35 WatchSource:0}: Error finding container 25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35: Status 404 returned error can't find the container with id 25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35 Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.034297 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.276208 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"078cd5e4-5781-4dc8-964e-6b1571ac85f9","Type":"ContainerStarted","Data":"c0fc7efead35042434f051962d9d9ac2dc4483ae4734dc3160267561cf55bd14"} Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.276472 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"078cd5e4-5781-4dc8-964e-6b1571ac85f9","Type":"ContainerStarted","Data":"25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35"} Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.278688 4708 generic.go:334] "Generic (PLEG): container finished" podID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerID="93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a" exitCode=0 Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.279480 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerDied","Data":"93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a"} Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.279505 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.279532 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerStarted","Data":"f77c4ddc10e530ad824cdec284fc0dfa01df60238b9589a3e5b3c0d7c1f01a30"} Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.287567 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=1.287547399 podStartE2EDuration="1.287547399s" podCreationTimestamp="2025-11-25 05:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:43:01.285460393 +0000 UTC m=+122.694293799" watchObservedRunningTime="2025-11-25 05:43:01.287547399 +0000 UTC m=+122.696380775" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.455750 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:43:01 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:43:01 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:43:01 crc kubenswrapper[4708]: healthz check failed Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.455802 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.508193 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.509574 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.512226 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.521960 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.653335 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.653439 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.653579 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pk9s\" (UniqueName: \"kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.757835 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pk9s\" (UniqueName: \"kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.758015 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.758240 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.758957 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.759259 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.779758 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pk9s\" (UniqueName: \"kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s\") pod \"redhat-operators-dvfr4\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.824513 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.907645 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.908608 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:01 crc kubenswrapper[4708]: I1125 05:43:01.923200 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.061739 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.061873 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.061920 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grg89\" (UniqueName: \"kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.162947 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.163017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.163049 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grg89\" (UniqueName: \"kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.163468 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.163702 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.179418 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grg89\" (UniqueName: \"kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89\") pod \"redhat-operators-bhvrv\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.225680 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.287012 4708 generic.go:334] "Generic (PLEG): container finished" podID="078cd5e4-5781-4dc8-964e-6b1571ac85f9" containerID="c0fc7efead35042434f051962d9d9ac2dc4483ae4734dc3160267561cf55bd14" exitCode=0 Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.287209 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"078cd5e4-5781-4dc8-964e-6b1571ac85f9","Type":"ContainerDied","Data":"c0fc7efead35042434f051962d9d9ac2dc4483ae4734dc3160267561cf55bd14"} Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.454238 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:43:02 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:43:02 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:43:02 crc kubenswrapper[4708]: healthz check failed Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.454311 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:43:02 crc kubenswrapper[4708]: I1125 05:43:02.917220 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qtrmv" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.034223 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.039547 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-qhzn8" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.142287 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.142326 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.158255 4708 patch_prober.go:28] interesting pod/console-f9d7485db-zkmtb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.158301 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zkmtb" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.455098 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.464828 4708 patch_prober.go:28] interesting pod/router-default-5444994796-r2zjk container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 05:43:03 crc kubenswrapper[4708]: [-]has-synced failed: reason withheld Nov 25 05:43:03 crc kubenswrapper[4708]: [+]process-running ok Nov 25 05:43:03 crc kubenswrapper[4708]: healthz check failed Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.464889 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r2zjk" podUID="db4f79f6-d226-40c5-bf33-e0f545525833" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.905729 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.907457 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.911293 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.911562 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 05:43:03 crc kubenswrapper[4708]: I1125 05:43:03.913358 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.101603 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.101650 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.203257 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.203300 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.203429 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.219348 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.232275 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.454896 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.457004 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r2zjk" Nov 25 05:43:04 crc kubenswrapper[4708]: I1125 05:43:04.875306 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.017110 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access\") pod \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.017208 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir\") pod \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\" (UID: \"078cd5e4-5781-4dc8-964e-6b1571ac85f9\") " Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.017406 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "078cd5e4-5781-4dc8-964e-6b1571ac85f9" (UID: "078cd5e4-5781-4dc8-964e-6b1571ac85f9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.022412 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "078cd5e4-5781-4dc8-964e-6b1571ac85f9" (UID: "078cd5e4-5781-4dc8-964e-6b1571ac85f9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.118189 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.118222 4708 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/078cd5e4-5781-4dc8-964e-6b1571ac85f9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.235891 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-r9f4b" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.327805 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.327796 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"078cd5e4-5781-4dc8-964e-6b1571ac85f9","Type":"ContainerDied","Data":"25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35"} Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.327856 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25c8543b336606815389acdac4f5dfc27cbb3a662c7d56d227aa411330614d35" Nov 25 05:43:05 crc kubenswrapper[4708]: I1125 05:43:05.975280 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:05 crc kubenswrapper[4708]: W1125 05:43:05.984255 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3ddd4b4_ab43_48fe_8c7d_c02311811cb8.slice/crio-9a26e0ab37f5c7d3747423869c2137fb00109b6e09cec54c0f13e0d82f224d08 WatchSource:0}: Error finding container 9a26e0ab37f5c7d3747423869c2137fb00109b6e09cec54c0f13e0d82f224d08: Status 404 returned error can't find the container with id 9a26e0ab37f5c7d3747423869c2137fb00109b6e09cec54c0f13e0d82f224d08 Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.034313 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.036031 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:06 crc kubenswrapper[4708]: W1125 05:43:06.043044 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf218c9c2_118a_4511_ba8c_277c70355fb2.slice/crio-b9a2a007c91969e4ed1490d87bbaf729d23ad802a2e47592d36605c65bcf0335 WatchSource:0}: Error finding container b9a2a007c91969e4ed1490d87bbaf729d23ad802a2e47592d36605c65bcf0335: Status 404 returned error can't find the container with id b9a2a007c91969e4ed1490d87bbaf729d23ad802a2e47592d36605c65bcf0335 Nov 25 05:43:06 crc kubenswrapper[4708]: W1125 05:43:06.047459 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod12ffe5c3_71d2_4363_9d67_612e080dfb79.slice/crio-2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939 WatchSource:0}: Error finding container 2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939: Status 404 returned error can't find the container with id 2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939 Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.056262 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.353441 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"12ffe5c3-71d2-4363-9d67-612e080dfb79","Type":"ContainerStarted","Data":"2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.356868 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerID="cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33" exitCode=0 Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.356926 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerDied","Data":"cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.356970 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerStarted","Data":"9a26e0ab37f5c7d3747423869c2137fb00109b6e09cec54c0f13e0d82f224d08"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.359891 4708 generic.go:334] "Generic (PLEG): container finished" podID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerID="a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3" exitCode=0 Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.359944 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerDied","Data":"a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.360066 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerStarted","Data":"b9a2a007c91969e4ed1490d87bbaf729d23ad802a2e47592d36605c65bcf0335"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.362961 4708 generic.go:334] "Generic (PLEG): container finished" podID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerID="8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c" exitCode=0 Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.362990 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerDied","Data":"8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c"} Nov 25 05:43:06 crc kubenswrapper[4708]: I1125 05:43:06.363007 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerStarted","Data":"02eca66e4dc5b85de8794c08d8f3333cd57506e36e93879d53904f3bfbf300df"} Nov 25 05:43:07 crc kubenswrapper[4708]: I1125 05:43:07.374902 4708 generic.go:334] "Generic (PLEG): container finished" podID="12ffe5c3-71d2-4363-9d67-612e080dfb79" containerID="620c0016d21e47b33a03a22568b1877fc310c975e80895c28252d1f76a86115c" exitCode=0 Nov 25 05:43:07 crc kubenswrapper[4708]: I1125 05:43:07.374967 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"12ffe5c3-71d2-4363-9d67-612e080dfb79","Type":"ContainerDied","Data":"620c0016d21e47b33a03a22568b1877fc310c975e80895c28252d1f76a86115c"} Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.145768 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.149322 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.659890 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.758848 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access\") pod \"12ffe5c3-71d2-4363-9d67-612e080dfb79\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.759052 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir\") pod \"12ffe5c3-71d2-4363-9d67-612e080dfb79\" (UID: \"12ffe5c3-71d2-4363-9d67-612e080dfb79\") " Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.759119 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "12ffe5c3-71d2-4363-9d67-612e080dfb79" (UID: "12ffe5c3-71d2-4363-9d67-612e080dfb79"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.759581 4708 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/12ffe5c3-71d2-4363-9d67-612e080dfb79-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.766863 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "12ffe5c3-71d2-4363-9d67-612e080dfb79" (UID: "12ffe5c3-71d2-4363-9d67-612e080dfb79"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:13 crc kubenswrapper[4708]: I1125 05:43:13.860640 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12ffe5c3-71d2-4363-9d67-612e080dfb79-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:14 crc kubenswrapper[4708]: I1125 05:43:14.425833 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"12ffe5c3-71d2-4363-9d67-612e080dfb79","Type":"ContainerDied","Data":"2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939"} Nov 25 05:43:14 crc kubenswrapper[4708]: I1125 05:43:14.425879 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f02522b62317e90949a21869186f237e5fb1aa3dbd4eab3ae3079f7ed9fd939" Nov 25 05:43:14 crc kubenswrapper[4708]: I1125 05:43:14.425929 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 05:43:17 crc kubenswrapper[4708]: I1125 05:43:17.440050 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerStarted","Data":"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.449779 4708 generic.go:334] "Generic (PLEG): container finished" podID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerID="6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.449852 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerDied","Data":"6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.451411 4708 generic.go:334] "Generic (PLEG): container finished" podID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerID="05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.451492 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerDied","Data":"05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.453512 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ca0d416-1d1f-49b7-869c-720144da0487" containerID="c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.453583 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerDied","Data":"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.455428 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerID="8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.455481 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerDied","Data":"8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.462781 4708 generic.go:334] "Generic (PLEG): container finished" podID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerID="a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.462835 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerDied","Data":"a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.467869 4708 generic.go:334] "Generic (PLEG): container finished" podID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerID="0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.467938 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerDied","Data":"0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.481981 4708 generic.go:334] "Generic (PLEG): container finished" podID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerID="85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.482210 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerDied","Data":"85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919"} Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.493684 4708 generic.go:334] "Generic (PLEG): container finished" podID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerID="48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe" exitCode=0 Nov 25 05:43:18 crc kubenswrapper[4708]: I1125 05:43:18.493721 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerDied","Data":"48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.323805 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.501039 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerStarted","Data":"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.504183 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerStarted","Data":"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.506198 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerStarted","Data":"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.508013 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerStarted","Data":"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.511431 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerStarted","Data":"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.513507 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerStarted","Data":"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.515571 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerStarted","Data":"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.517266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerStarted","Data":"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484"} Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.524924 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sjf6r" podStartSLOduration=2.851510534 podStartE2EDuration="21.524913647s" podCreationTimestamp="2025-11-25 05:42:58 +0000 UTC" firstStartedPulling="2025-11-25 05:43:00.265673272 +0000 UTC m=+121.674506648" lastFinishedPulling="2025-11-25 05:43:18.939076365 +0000 UTC m=+140.347909761" observedRunningTime="2025-11-25 05:43:19.523064321 +0000 UTC m=+140.931897697" watchObservedRunningTime="2025-11-25 05:43:19.524913647 +0000 UTC m=+140.933747032" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.545488 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m5q65" podStartSLOduration=2.849222548 podStartE2EDuration="21.545477811s" podCreationTimestamp="2025-11-25 05:42:58 +0000 UTC" firstStartedPulling="2025-11-25 05:43:00.25507774 +0000 UTC m=+121.663911125" lastFinishedPulling="2025-11-25 05:43:18.951333002 +0000 UTC m=+140.360166388" observedRunningTime="2025-11-25 05:43:19.544137274 +0000 UTC m=+140.952970661" watchObservedRunningTime="2025-11-25 05:43:19.545477811 +0000 UTC m=+140.954311197" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.565650 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fz4gp" podStartSLOduration=1.843085104 podStartE2EDuration="21.565628216s" podCreationTimestamp="2025-11-25 05:42:58 +0000 UTC" firstStartedPulling="2025-11-25 05:42:59.237792684 +0000 UTC m=+120.646626070" lastFinishedPulling="2025-11-25 05:43:18.960335796 +0000 UTC m=+140.369169182" observedRunningTime="2025-11-25 05:43:19.561331976 +0000 UTC m=+140.970165362" watchObservedRunningTime="2025-11-25 05:43:19.565628216 +0000 UTC m=+140.974461813" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.578172 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dvfr4" podStartSLOduration=5.991679618 podStartE2EDuration="18.578152313s" podCreationTimestamp="2025-11-25 05:43:01 +0000 UTC" firstStartedPulling="2025-11-25 05:43:06.364555285 +0000 UTC m=+127.773388671" lastFinishedPulling="2025-11-25 05:43:18.95102798 +0000 UTC m=+140.359861366" observedRunningTime="2025-11-25 05:43:19.576218897 +0000 UTC m=+140.985052273" watchObservedRunningTime="2025-11-25 05:43:19.578152313 +0000 UTC m=+140.986985689" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.594275 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x6m7c" podStartSLOduration=5.37087441 podStartE2EDuration="19.594247367s" podCreationTimestamp="2025-11-25 05:43:00 +0000 UTC" firstStartedPulling="2025-11-25 05:43:04.843028808 +0000 UTC m=+126.251862194" lastFinishedPulling="2025-11-25 05:43:19.066401765 +0000 UTC m=+140.475235151" observedRunningTime="2025-11-25 05:43:19.592641256 +0000 UTC m=+141.001474642" watchObservedRunningTime="2025-11-25 05:43:19.594247367 +0000 UTC m=+141.003080753" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.611220 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h47dj" podStartSLOduration=6.983584778 podStartE2EDuration="19.611194547s" podCreationTimestamp="2025-11-25 05:43:00 +0000 UTC" firstStartedPulling="2025-11-25 05:43:06.361699212 +0000 UTC m=+127.770532598" lastFinishedPulling="2025-11-25 05:43:18.989308982 +0000 UTC m=+140.398142367" observedRunningTime="2025-11-25 05:43:19.609541826 +0000 UTC m=+141.018375212" watchObservedRunningTime="2025-11-25 05:43:19.611194547 +0000 UTC m=+141.020027933" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.642986 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bhvrv" podStartSLOduration=6.079860017 podStartE2EDuration="18.642967721s" podCreationTimestamp="2025-11-25 05:43:01 +0000 UTC" firstStartedPulling="2025-11-25 05:43:06.359171718 +0000 UTC m=+127.768005104" lastFinishedPulling="2025-11-25 05:43:18.922279422 +0000 UTC m=+140.331112808" observedRunningTime="2025-11-25 05:43:19.625441287 +0000 UTC m=+141.034274673" watchObservedRunningTime="2025-11-25 05:43:19.642967721 +0000 UTC m=+141.051801107" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.645473 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xfh2t" podStartSLOduration=2.8527025569999998 podStartE2EDuration="21.645465342s" podCreationTimestamp="2025-11-25 05:42:58 +0000 UTC" firstStartedPulling="2025-11-25 05:43:00.266744066 +0000 UTC m=+121.675577442" lastFinishedPulling="2025-11-25 05:43:19.05950684 +0000 UTC m=+140.468340227" observedRunningTime="2025-11-25 05:43:19.643812291 +0000 UTC m=+141.052645677" watchObservedRunningTime="2025-11-25 05:43:19.645465342 +0000 UTC m=+141.054298728" Nov 25 05:43:19 crc kubenswrapper[4708]: I1125 05:43:19.958444 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:43:20 crc kubenswrapper[4708]: I1125 05:43:20.624886 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:20 crc kubenswrapper[4708]: I1125 05:43:20.625138 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:20 crc kubenswrapper[4708]: I1125 05:43:20.711203 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:21 crc kubenswrapper[4708]: I1125 05:43:21.035101 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:21 crc kubenswrapper[4708]: I1125 05:43:21.035274 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:21 crc kubenswrapper[4708]: I1125 05:43:21.071725 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:21 crc kubenswrapper[4708]: I1125 05:43:21.824843 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:21 crc kubenswrapper[4708]: I1125 05:43:21.824902 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:22 crc kubenswrapper[4708]: I1125 05:43:22.226204 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:22 crc kubenswrapper[4708]: I1125 05:43:22.226257 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:22 crc kubenswrapper[4708]: I1125 05:43:22.860234 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dvfr4" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="registry-server" probeResult="failure" output=< Nov 25 05:43:22 crc kubenswrapper[4708]: timeout: failed to connect service ":50051" within 1s Nov 25 05:43:22 crc kubenswrapper[4708]: > Nov 25 05:43:23 crc kubenswrapper[4708]: I1125 05:43:23.262072 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bhvrv" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="registry-server" probeResult="failure" output=< Nov 25 05:43:23 crc kubenswrapper[4708]: timeout: failed to connect service ":50051" within 1s Nov 25 05:43:23 crc kubenswrapper[4708]: > Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.911948 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.912878 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.913005 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.913099 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.915608 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.915719 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.915952 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.924462 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.925015 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.930084 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.936585 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:43:24 crc kubenswrapper[4708]: I1125 05:43:24.936706 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:43:25 crc kubenswrapper[4708]: I1125 05:43:25.108667 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 05:43:25 crc kubenswrapper[4708]: I1125 05:43:25.112893 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 05:43:25 crc kubenswrapper[4708]: I1125 05:43:25.209485 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:43:25 crc kubenswrapper[4708]: I1125 05:43:25.560447 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"14cd1028126f28ba75c5dee6a69dc6950982d560ff08025973a976d466e1f9c1"} Nov 25 05:43:25 crc kubenswrapper[4708]: W1125 05:43:25.564898 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-54a0b149320c6d88e2317c2647e74f6b7d7d1a302a07d33061cdccdad53d929b WatchSource:0}: Error finding container 54a0b149320c6d88e2317c2647e74f6b7d7d1a302a07d33061cdccdad53d929b: Status 404 returned error can't find the container with id 54a0b149320c6d88e2317c2647e74f6b7d7d1a302a07d33061cdccdad53d929b Nov 25 05:43:25 crc kubenswrapper[4708]: W1125 05:43:25.637733 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-4240a8470bd9c5d18c46ecaecc8bde58137d6606f635329db5529ffea241aba0 WatchSource:0}: Error finding container 4240a8470bd9c5d18c46ecaecc8bde58137d6606f635329db5529ffea241aba0: Status 404 returned error can't find the container with id 4240a8470bd9c5d18c46ecaecc8bde58137d6606f635329db5529ffea241aba0 Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.566561 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"06d46e38126a2cfbc35fcc21800a2885916a4eee14e706749ba12408f38cec5f"} Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.567268 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"54a0b149320c6d88e2317c2647e74f6b7d7d1a302a07d33061cdccdad53d929b"} Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.567829 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"08ca058743a2e9590596e61451fb0c03d4ac93609e4186848a1ba370cb694d8c"} Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.569707 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f008d415330ded30697a7719d34b24f61f82f7ec6bb6c69e4a4c8f03c638d42b"} Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.569739 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4240a8470bd9c5d18c46ecaecc8bde58137d6606f635329db5529ffea241aba0"} Nov 25 05:43:26 crc kubenswrapper[4708]: I1125 05:43:26.569919 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.824401 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.825351 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.862653 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.945172 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.945404 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:43:28 crc kubenswrapper[4708]: I1125 05:43:28.974431 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.026122 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.026175 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.055445 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.256865 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.256913 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.291106 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.629571 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.629641 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.634811 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:29 crc kubenswrapper[4708]: I1125 05:43:29.637601 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:43:30 crc kubenswrapper[4708]: I1125 05:43:30.657174 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:43:30 crc kubenswrapper[4708]: I1125 05:43:30.712967 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.065363 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.313685 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.600439 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xfh2t" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="registry-server" containerID="cri-o://709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc" gracePeriod=2 Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.601052 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sjf6r" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="registry-server" containerID="cri-o://1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae" gracePeriod=2 Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.858534 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:31 crc kubenswrapper[4708]: I1125 05:43:31.890349 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.056688 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.095013 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities\") pod \"3ca0d416-1d1f-49b7-869c-720144da0487\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.095343 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content\") pod \"3ca0d416-1d1f-49b7-869c-720144da0487\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.095411 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7r5pb\" (UniqueName: \"kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb\") pod \"3ca0d416-1d1f-49b7-869c-720144da0487\" (UID: \"3ca0d416-1d1f-49b7-869c-720144da0487\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.095813 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities" (OuterVolumeSpecName: "utilities") pod "3ca0d416-1d1f-49b7-869c-720144da0487" (UID: "3ca0d416-1d1f-49b7-869c-720144da0487"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.096066 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.102700 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb" (OuterVolumeSpecName: "kube-api-access-7r5pb") pod "3ca0d416-1d1f-49b7-869c-720144da0487" (UID: "3ca0d416-1d1f-49b7-869c-720144da0487"). InnerVolumeSpecName "kube-api-access-7r5pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.124450 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.135700 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3ca0d416-1d1f-49b7-869c-720144da0487" (UID: "3ca0d416-1d1f-49b7-869c-720144da0487"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197010 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcwzq\" (UniqueName: \"kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq\") pod \"c1631d73-34ab-4e9b-8261-a91bcabd762b\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197086 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities\") pod \"c1631d73-34ab-4e9b-8261-a91bcabd762b\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197221 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content\") pod \"c1631d73-34ab-4e9b-8261-a91bcabd762b\" (UID: \"c1631d73-34ab-4e9b-8261-a91bcabd762b\") " Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197552 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ca0d416-1d1f-49b7-869c-720144da0487-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197572 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7r5pb\" (UniqueName: \"kubernetes.io/projected/3ca0d416-1d1f-49b7-869c-720144da0487-kube-api-access-7r5pb\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.197791 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities" (OuterVolumeSpecName: "utilities") pod "c1631d73-34ab-4e9b-8261-a91bcabd762b" (UID: "c1631d73-34ab-4e9b-8261-a91bcabd762b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.199424 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq" (OuterVolumeSpecName: "kube-api-access-dcwzq") pod "c1631d73-34ab-4e9b-8261-a91bcabd762b" (UID: "c1631d73-34ab-4e9b-8261-a91bcabd762b"). InnerVolumeSpecName "kube-api-access-dcwzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.236599 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1631d73-34ab-4e9b-8261-a91bcabd762b" (UID: "c1631d73-34ab-4e9b-8261-a91bcabd762b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.257886 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.287290 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.298411 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.298439 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1631d73-34ab-4e9b-8261-a91bcabd762b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.298451 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcwzq\" (UniqueName: \"kubernetes.io/projected/c1631d73-34ab-4e9b-8261-a91bcabd762b-kube-api-access-dcwzq\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.608137 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ca0d416-1d1f-49b7-869c-720144da0487" containerID="1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae" exitCode=0 Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.608203 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerDied","Data":"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae"} Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.608301 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sjf6r" event={"ID":"3ca0d416-1d1f-49b7-869c-720144da0487","Type":"ContainerDied","Data":"fc8c7c55b4dd7147d28e23b59994d6998fa87b11fdf951dd9e08ea22042dba75"} Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.608332 4708 scope.go:117] "RemoveContainer" containerID="1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.608410 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sjf6r" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.610610 4708 generic.go:334] "Generic (PLEG): container finished" podID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerID="709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc" exitCode=0 Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.610657 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xfh2t" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.610736 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerDied","Data":"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc"} Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.610772 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xfh2t" event={"ID":"c1631d73-34ab-4e9b-8261-a91bcabd762b","Type":"ContainerDied","Data":"e8566f742cc1c83f9f444b7290bae8d5efb62a4386d998545d130c52c2b35cb5"} Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.622044 4708 scope.go:117] "RemoveContainer" containerID="c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.638914 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.640872 4708 scope.go:117] "RemoveContainer" containerID="4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.641686 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sjf6r"] Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.648316 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.650469 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xfh2t"] Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.654079 4708 scope.go:117] "RemoveContainer" containerID="1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.654455 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae\": container with ID starting with 1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae not found: ID does not exist" containerID="1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.654495 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae"} err="failed to get container status \"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae\": rpc error: code = NotFound desc = could not find container \"1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae\": container with ID starting with 1c4697acde10ab29b82fcb579ecd25f10a6a3dcf5f81f7532a11016deb4cceae not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.654549 4708 scope.go:117] "RemoveContainer" containerID="c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.654782 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1\": container with ID starting with c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1 not found: ID does not exist" containerID="c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.654805 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1"} err="failed to get container status \"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1\": rpc error: code = NotFound desc = could not find container \"c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1\": container with ID starting with c364e09670776a8341f6646d35bf0dcdac6cb3a9eb2e630b21daf9ee538ca8f1 not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.654817 4708 scope.go:117] "RemoveContainer" containerID="4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.655079 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7\": container with ID starting with 4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7 not found: ID does not exist" containerID="4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.655125 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7"} err="failed to get container status \"4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7\": rpc error: code = NotFound desc = could not find container \"4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7\": container with ID starting with 4d6d951a3444b46bf4b2ac311e17752440eb006ec97f4f1790aeb027205703b7 not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.655150 4708 scope.go:117] "RemoveContainer" containerID="709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.664345 4708 scope.go:117] "RemoveContainer" containerID="85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.674780 4708 scope.go:117] "RemoveContainer" containerID="80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686038 4708 scope.go:117] "RemoveContainer" containerID="709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.686342 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc\": container with ID starting with 709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc not found: ID does not exist" containerID="709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686370 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc"} err="failed to get container status \"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc\": rpc error: code = NotFound desc = could not find container \"709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc\": container with ID starting with 709dff689ebd5f56d6a9f49752450ac9a8f275bde7389825013af8da9faa7abc not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686387 4708 scope.go:117] "RemoveContainer" containerID="85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.686659 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919\": container with ID starting with 85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919 not found: ID does not exist" containerID="85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686689 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919"} err="failed to get container status \"85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919\": rpc error: code = NotFound desc = could not find container \"85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919\": container with ID starting with 85ee9664c14bc8a8a4698f26ec7d8d6ea02aa0da45716e129dd213df10662919 not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686710 4708 scope.go:117] "RemoveContainer" containerID="80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68" Nov 25 05:43:32 crc kubenswrapper[4708]: E1125 05:43:32.686940 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68\": container with ID starting with 80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68 not found: ID does not exist" containerID="80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.686966 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68"} err="failed to get container status \"80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68\": rpc error: code = NotFound desc = could not find container \"80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68\": container with ID starting with 80b27983897ebb858f7d78e55cfc5470563cacbe454c341104ecc1a9c4c7ee68 not found: ID does not exist" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.900077 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" path="/var/lib/kubelet/pods/3ca0d416-1d1f-49b7-869c-720144da0487/volumes" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.900921 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" path="/var/lib/kubelet/pods/c1631d73-34ab-4e9b-8261-a91bcabd762b/volumes" Nov 25 05:43:32 crc kubenswrapper[4708]: I1125 05:43:32.920200 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.112867 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.113335 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h47dj" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="registry-server" containerID="cri-o://6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e" gracePeriod=2 Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.517303 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-wd8m8" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.534315 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.614470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xxzr\" (UniqueName: \"kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr\") pod \"f218c9c2-118a-4511-ba8c-277c70355fb2\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.614588 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities\") pod \"f218c9c2-118a-4511-ba8c-277c70355fb2\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.614753 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content\") pod \"f218c9c2-118a-4511-ba8c-277c70355fb2\" (UID: \"f218c9c2-118a-4511-ba8c-277c70355fb2\") " Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.615419 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities" (OuterVolumeSpecName: "utilities") pod "f218c9c2-118a-4511-ba8c-277c70355fb2" (UID: "f218c9c2-118a-4511-ba8c-277c70355fb2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.616408 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621207 4708 generic.go:334] "Generic (PLEG): container finished" podID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerID="6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e" exitCode=0 Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621324 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr" (OuterVolumeSpecName: "kube-api-access-5xxzr") pod "f218c9c2-118a-4511-ba8c-277c70355fb2" (UID: "f218c9c2-118a-4511-ba8c-277c70355fb2"). InnerVolumeSpecName "kube-api-access-5xxzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621448 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h47dj" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621474 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerDied","Data":"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e"} Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621539 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h47dj" event={"ID":"f218c9c2-118a-4511-ba8c-277c70355fb2","Type":"ContainerDied","Data":"b9a2a007c91969e4ed1490d87bbaf729d23ad802a2e47592d36605c65bcf0335"} Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.621559 4708 scope.go:117] "RemoveContainer" containerID="6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.627832 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f218c9c2-118a-4511-ba8c-277c70355fb2" (UID: "f218c9c2-118a-4511-ba8c-277c70355fb2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.645636 4708 scope.go:117] "RemoveContainer" containerID="a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.661476 4708 scope.go:117] "RemoveContainer" containerID="a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.677393 4708 scope.go:117] "RemoveContainer" containerID="6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e" Nov 25 05:43:33 crc kubenswrapper[4708]: E1125 05:43:33.677795 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e\": container with ID starting with 6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e not found: ID does not exist" containerID="6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.677832 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e"} err="failed to get container status \"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e\": rpc error: code = NotFound desc = could not find container \"6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e\": container with ID starting with 6bd8c30528199a911cfa84c7988bd8bb8c37e96c500b480ca533fc049fbcb33e not found: ID does not exist" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.677861 4708 scope.go:117] "RemoveContainer" containerID="a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228" Nov 25 05:43:33 crc kubenswrapper[4708]: E1125 05:43:33.678123 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228\": container with ID starting with a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228 not found: ID does not exist" containerID="a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.678155 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228"} err="failed to get container status \"a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228\": rpc error: code = NotFound desc = could not find container \"a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228\": container with ID starting with a249ea5db193d0fd0b9a2953ae5c3ba5ab1efebac7a698cd856f4864e7811228 not found: ID does not exist" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.678175 4708 scope.go:117] "RemoveContainer" containerID="a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3" Nov 25 05:43:33 crc kubenswrapper[4708]: E1125 05:43:33.678396 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3\": container with ID starting with a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3 not found: ID does not exist" containerID="a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.678421 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3"} err="failed to get container status \"a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3\": rpc error: code = NotFound desc = could not find container \"a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3\": container with ID starting with a5bae065497fb3f5098923eb75fea180c8587b119818e99977fc43dbfa4538a3 not found: ID does not exist" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.717347 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f218c9c2-118a-4511-ba8c-277c70355fb2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.717374 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xxzr\" (UniqueName: \"kubernetes.io/projected/f218c9c2-118a-4511-ba8c-277c70355fb2-kube-api-access-5xxzr\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.948703 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:33 crc kubenswrapper[4708]: I1125 05:43:33.954049 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h47dj"] Nov 25 05:43:34 crc kubenswrapper[4708]: I1125 05:43:34.898210 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" path="/var/lib/kubelet/pods/f218c9c2-118a-4511-ba8c-277c70355fb2/volumes" Nov 25 05:43:35 crc kubenswrapper[4708]: I1125 05:43:35.711104 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:35 crc kubenswrapper[4708]: I1125 05:43:35.711754 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bhvrv" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="registry-server" containerID="cri-o://fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40" gracePeriod=2 Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.097591 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.242780 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grg89\" (UniqueName: \"kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89\") pod \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.242893 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content\") pod \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.242918 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities\") pod \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\" (UID: \"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8\") " Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.243692 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities" (OuterVolumeSpecName: "utilities") pod "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" (UID: "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.247360 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89" (OuterVolumeSpecName: "kube-api-access-grg89") pod "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" (UID: "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8"). InnerVolumeSpecName "kube-api-access-grg89". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.307217 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" (UID: "b3ddd4b4-ab43-48fe-8c7d-c02311811cb8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.344581 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.344602 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.344612 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grg89\" (UniqueName: \"kubernetes.io/projected/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8-kube-api-access-grg89\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.640598 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerID="fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40" exitCode=0 Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.640642 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerDied","Data":"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40"} Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.640680 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bhvrv" event={"ID":"b3ddd4b4-ab43-48fe-8c7d-c02311811cb8","Type":"ContainerDied","Data":"9a26e0ab37f5c7d3747423869c2137fb00109b6e09cec54c0f13e0d82f224d08"} Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.640701 4708 scope.go:117] "RemoveContainer" containerID="fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.640815 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bhvrv" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.654931 4708 scope.go:117] "RemoveContainer" containerID="8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.663427 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.665415 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bhvrv"] Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.671020 4708 scope.go:117] "RemoveContainer" containerID="cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.690042 4708 scope.go:117] "RemoveContainer" containerID="fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40" Nov 25 05:43:36 crc kubenswrapper[4708]: E1125 05:43:36.690513 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40\": container with ID starting with fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40 not found: ID does not exist" containerID="fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.690554 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40"} err="failed to get container status \"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40\": rpc error: code = NotFound desc = could not find container \"fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40\": container with ID starting with fa2e6e261c9f668664222e53219a8f0367b5dd2f4547f9207ee1f226d2b93f40 not found: ID does not exist" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.690576 4708 scope.go:117] "RemoveContainer" containerID="8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d" Nov 25 05:43:36 crc kubenswrapper[4708]: E1125 05:43:36.690898 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d\": container with ID starting with 8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d not found: ID does not exist" containerID="8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.690921 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d"} err="failed to get container status \"8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d\": rpc error: code = NotFound desc = could not find container \"8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d\": container with ID starting with 8ea49e84c5e3119b87d1bd414c798a58cfb257697f5d1cf768a7316b52df1d3d not found: ID does not exist" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.690936 4708 scope.go:117] "RemoveContainer" containerID="cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33" Nov 25 05:43:36 crc kubenswrapper[4708]: E1125 05:43:36.691227 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33\": container with ID starting with cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33 not found: ID does not exist" containerID="cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.691256 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33"} err="failed to get container status \"cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33\": rpc error: code = NotFound desc = could not find container \"cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33\": container with ID starting with cf7117071187c4a009c56e16ba1a9fde6b4ea5f1a7e32985c66beedde03bbc33 not found: ID does not exist" Nov 25 05:43:36 crc kubenswrapper[4708]: I1125 05:43:36.901968 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" path="/var/lib/kubelet/pods/b3ddd4b4-ab43-48fe-8c7d-c02311811cb8/volumes" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.166030 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.167489 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.181934 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/711da2a4-b5f2-40ce-84d4-619c4e3f2522-metrics-certs\") pod \"network-metrics-daemon-g6lks\" (UID: \"711da2a4-b5f2-40ce-84d4-619c4e3f2522\") " pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.406183 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.414496 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-g6lks" Nov 25 05:43:38 crc kubenswrapper[4708]: I1125 05:43:38.823401 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-g6lks"] Nov 25 05:43:39 crc kubenswrapper[4708]: I1125 05:43:39.660273 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-g6lks" event={"ID":"711da2a4-b5f2-40ce-84d4-619c4e3f2522","Type":"ContainerStarted","Data":"a8231e0f49f0e6b5774181c1b99de7e649fb5a84fffc0419fd2b699d127a8f82"} Nov 25 05:43:39 crc kubenswrapper[4708]: I1125 05:43:39.660564 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-g6lks" event={"ID":"711da2a4-b5f2-40ce-84d4-619c4e3f2522","Type":"ContainerStarted","Data":"ac565c0c68de8dcfe1ccc06215f88f08b94ded4854e2cfd38695545053ca95c2"} Nov 25 05:43:39 crc kubenswrapper[4708]: I1125 05:43:39.660578 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-g6lks" event={"ID":"711da2a4-b5f2-40ce-84d4-619c4e3f2522","Type":"ContainerStarted","Data":"21f84ff693da32dc26aa59731e9970b5e1ddd0a0389a0e632acb96f39117a852"} Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.939684 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-g6lks" podStartSLOduration=145.939665742 podStartE2EDuration="2m25.939665742s" podCreationTimestamp="2025-11-25 05:41:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:43:39.677481256 +0000 UTC m=+161.086314642" watchObservedRunningTime="2025-11-25 05:43:41.939665742 +0000 UTC m=+163.348499128" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.942326 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.942581 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.942665 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.942727 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="078cd5e4-5781-4dc8-964e-6b1571ac85f9" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.942776 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="078cd5e4-5781-4dc8-964e-6b1571ac85f9" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.942821 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.942864 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.942907 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.942955 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943003 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943054 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943103 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943147 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943190 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943237 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943287 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943355 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943405 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12ffe5c3-71d2-4363-9d67-612e080dfb79" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943448 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="12ffe5c3-71d2-4363-9d67-612e080dfb79" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943492 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943554 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943614 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943660 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943705 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943753 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="extract-content" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943804 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943848 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: E1125 05:43:41.943897 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.943945 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="extract-utilities" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944064 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="12ffe5c3-71d2-4363-9d67-612e080dfb79" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944117 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1631d73-34ab-4e9b-8261-a91bcabd762b" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944179 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="078cd5e4-5781-4dc8-964e-6b1571ac85f9" containerName="pruner" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944231 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f218c9c2-118a-4511-ba8c-277c70355fb2" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944276 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3ddd4b4-ab43-48fe-8c7d-c02311811cb8" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944338 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ca0d416-1d1f-49b7-869c-720144da0487" containerName="registry-server" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.944714 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.946506 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.946509 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 05:43:41 crc kubenswrapper[4708]: I1125 05:43:41.952846 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.113937 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.114145 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.215209 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.215468 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.215620 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.230713 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.258611 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.621982 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 05:43:42 crc kubenswrapper[4708]: I1125 05:43:42.677083 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4973b840-b5c6-4cf1-8054-e9e535adeaaf","Type":"ContainerStarted","Data":"9bbe7d7e747dc7db88c044b3fe6a484b6826842f03ead016c42013c76a44c6e2"} Nov 25 05:43:43 crc kubenswrapper[4708]: I1125 05:43:43.685249 4708 generic.go:334] "Generic (PLEG): container finished" podID="4973b840-b5c6-4cf1-8054-e9e535adeaaf" containerID="83c24e7fa295b9a2a8c8eea351e47c0c45907d529cf3e7598b50f7f89fa4f60d" exitCode=0 Nov 25 05:43:43 crc kubenswrapper[4708]: I1125 05:43:43.685365 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4973b840-b5c6-4cf1-8054-e9e535adeaaf","Type":"ContainerDied","Data":"83c24e7fa295b9a2a8c8eea351e47c0c45907d529cf3e7598b50f7f89fa4f60d"} Nov 25 05:43:44 crc kubenswrapper[4708]: I1125 05:43:44.356974 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:43:44 crc kubenswrapper[4708]: I1125 05:43:44.357068 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:43:44 crc kubenswrapper[4708]: I1125 05:43:44.996694 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.151627 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir\") pod \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.151790 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4973b840-b5c6-4cf1-8054-e9e535adeaaf" (UID: "4973b840-b5c6-4cf1-8054-e9e535adeaaf"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.151843 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access\") pod \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\" (UID: \"4973b840-b5c6-4cf1-8054-e9e535adeaaf\") " Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.152106 4708 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.157429 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4973b840-b5c6-4cf1-8054-e9e535adeaaf" (UID: "4973b840-b5c6-4cf1-8054-e9e535adeaaf"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.253738 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4973b840-b5c6-4cf1-8054-e9e535adeaaf-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.695169 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"4973b840-b5c6-4cf1-8054-e9e535adeaaf","Type":"ContainerDied","Data":"9bbe7d7e747dc7db88c044b3fe6a484b6826842f03ead016c42013c76a44c6e2"} Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.695205 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bbe7d7e747dc7db88c044b3fe6a484b6826842f03ead016c42013c76a44c6e2" Nov 25 05:43:45 crc kubenswrapper[4708]: I1125 05:43:45.695245 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.737399 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 05:43:48 crc kubenswrapper[4708]: E1125 05:43:48.737842 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4973b840-b5c6-4cf1-8054-e9e535adeaaf" containerName="pruner" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.737854 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4973b840-b5c6-4cf1-8054-e9e535adeaaf" containerName="pruner" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.737962 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4973b840-b5c6-4cf1-8054-e9e535adeaaf" containerName="pruner" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.738348 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.743042 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.743542 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.745546 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.895288 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.895352 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.895370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.996351 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.996395 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.996417 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.996541 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:48 crc kubenswrapper[4708]: I1125 05:43:48.996578 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.012710 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access\") pod \"installer-9-crc\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.053942 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.414584 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.717452 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7f521c0f-a241-4990-b984-ac475119e8d8","Type":"ContainerStarted","Data":"7c01b3977b7679972a415340e75baaae1074324090e5d0abe8ce38816c0efbea"} Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.717828 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7f521c0f-a241-4990-b984-ac475119e8d8","Type":"ContainerStarted","Data":"f61d7477f4eab755ef27109a4dfdabc7689e75117e2f414255aca6d1b24113f5"} Nov 25 05:43:49 crc kubenswrapper[4708]: I1125 05:43:49.730155 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.730139394 podStartE2EDuration="1.730139394s" podCreationTimestamp="2025-11-25 05:43:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:43:49.728590703 +0000 UTC m=+171.137424088" watchObservedRunningTime="2025-11-25 05:43:49.730139394 +0000 UTC m=+171.138972781" Nov 25 05:43:55 crc kubenswrapper[4708]: I1125 05:43:55.215073 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 05:43:57 crc kubenswrapper[4708]: I1125 05:43:57.940928 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerName="oauth-openshift" containerID="cri-o://c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc" gracePeriod=15 Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.268968 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.296482 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-df6677757-j5gpm"] Nov 25 05:43:58 crc kubenswrapper[4708]: E1125 05:43:58.296763 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerName="oauth-openshift" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.296777 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerName="oauth-openshift" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.296877 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerName="oauth-openshift" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.297290 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.309236 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-df6677757-j5gpm"] Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.404724 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405071 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lf2vq\" (UniqueName: \"kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405161 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405190 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405209 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405250 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405282 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405320 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405382 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405413 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405455 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405476 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405497 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405557 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login\") pod \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\" (UID: \"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c\") " Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405548 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405752 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-router-certs\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405775 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405799 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405820 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-service-ca\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405838 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-error\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405861 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-session\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405882 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405898 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-policies\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405918 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ktbp\" (UniqueName: \"kubernetes.io/projected/9bf1834b-f548-4aad-994c-1ffa5c212eda-kube-api-access-9ktbp\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405937 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-dir\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405973 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.405992 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-login\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.406016 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.406038 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.406072 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.406685 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.406919 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.407314 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.407756 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.412171 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq" (OuterVolumeSpecName: "kube-api-access-lf2vq") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "kube-api-access-lf2vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.412618 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.412886 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413114 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413319 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413483 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413642 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413744 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.413918 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" (UID: "3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.507229 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-service-ca\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.507870 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-service-ca\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.507937 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-error\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.507969 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-session\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.507991 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508022 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-policies\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508049 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ktbp\" (UniqueName: \"kubernetes.io/projected/9bf1834b-f548-4aad-994c-1ffa5c212eda-kube-api-access-9ktbp\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508076 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-dir\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508119 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508141 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-login\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508165 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508188 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508218 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-router-certs\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508235 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508256 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508293 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508304 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508314 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508323 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508334 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508343 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508358 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508367 4708 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508375 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508386 4708 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508396 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508406 4708 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.508418 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lf2vq\" (UniqueName: \"kubernetes.io/projected/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c-kube-api-access-lf2vq\") on node \"crc\" DevicePath \"\"" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.509322 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-policies\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.509598 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9bf1834b-f548-4aad-994c-1ffa5c212eda-audit-dir\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.509841 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-cliconfig\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.510076 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-error\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.511077 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.511500 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-session\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.512097 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.512098 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.512476 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-serving-cert\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.512574 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.512751 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-user-template-login\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.513982 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9bf1834b-f548-4aad-994c-1ffa5c212eda-v4-0-config-system-router-certs\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.523890 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ktbp\" (UniqueName: \"kubernetes.io/projected/9bf1834b-f548-4aad-994c-1ffa5c212eda-kube-api-access-9ktbp\") pod \"oauth-openshift-df6677757-j5gpm\" (UID: \"9bf1834b-f548-4aad-994c-1ffa5c212eda\") " pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.613764 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.772312 4708 generic.go:334] "Generic (PLEG): container finished" podID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" containerID="c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc" exitCode=0 Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.772355 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" event={"ID":"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c","Type":"ContainerDied","Data":"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc"} Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.772381 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" event={"ID":"3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c","Type":"ContainerDied","Data":"820cf8151b677b0081ef1324a393dd5bca3e3725c3638abfe287cff77d9abce6"} Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.772397 4708 scope.go:117] "RemoveContainer" containerID="c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.772503 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-2kwv6" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.778578 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-df6677757-j5gpm"] Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.788237 4708 scope.go:117] "RemoveContainer" containerID="c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc" Nov 25 05:43:58 crc kubenswrapper[4708]: E1125 05:43:58.788653 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc\": container with ID starting with c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc not found: ID does not exist" containerID="c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc" Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.788689 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc"} err="failed to get container status \"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc\": rpc error: code = NotFound desc = could not find container \"c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc\": container with ID starting with c4b151f03afc115b3268222cfd9aca52c7cd682e0004f1a87732119c6cca00bc not found: ID does not exist" Nov 25 05:43:58 crc kubenswrapper[4708]: W1125 05:43:58.792813 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bf1834b_f548_4aad_994c_1ffa5c212eda.slice/crio-8e9cce19c3a469f06c96d3775dbb481641dfaee60c41f865367010f3c5e74c44 WatchSource:0}: Error finding container 8e9cce19c3a469f06c96d3775dbb481641dfaee60c41f865367010f3c5e74c44: Status 404 returned error can't find the container with id 8e9cce19c3a469f06c96d3775dbb481641dfaee60c41f865367010f3c5e74c44 Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.799131 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.801296 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-2kwv6"] Nov 25 05:43:58 crc kubenswrapper[4708]: I1125 05:43:58.917921 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c" path="/var/lib/kubelet/pods/3f7427b2-e742-48fb-8f6f-f6f6b6c1cd5c/volumes" Nov 25 05:43:59 crc kubenswrapper[4708]: I1125 05:43:59.780335 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" event={"ID":"9bf1834b-f548-4aad-994c-1ffa5c212eda","Type":"ContainerStarted","Data":"55e9842be8838113465463a2fa7268e309425435f8321c30b9afa7d47623c8a5"} Nov 25 05:43:59 crc kubenswrapper[4708]: I1125 05:43:59.780734 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" event={"ID":"9bf1834b-f548-4aad-994c-1ffa5c212eda","Type":"ContainerStarted","Data":"8e9cce19c3a469f06c96d3775dbb481641dfaee60c41f865367010f3c5e74c44"} Nov 25 05:43:59 crc kubenswrapper[4708]: I1125 05:43:59.780932 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:59 crc kubenswrapper[4708]: I1125 05:43:59.789729 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" Nov 25 05:43:59 crc kubenswrapper[4708]: I1125 05:43:59.800182 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-df6677757-j5gpm" podStartSLOduration=27.800160375 podStartE2EDuration="27.800160375s" podCreationTimestamp="2025-11-25 05:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:43:59.796671925 +0000 UTC m=+181.205505311" watchObservedRunningTime="2025-11-25 05:43:59.800160375 +0000 UTC m=+181.208993761" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.772489 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.773234 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m5q65" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="registry-server" containerID="cri-o://249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484" gracePeriod=30 Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.799980 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.800790 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fz4gp" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="registry-server" containerID="cri-o://2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e" gracePeriod=30 Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.806280 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.810594 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerName="marketplace-operator" containerID="cri-o://7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab" gracePeriod=30 Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.814174 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.814393 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x6m7c" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="registry-server" containerID="cri-o://589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c" gracePeriod=30 Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.816422 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlzlc"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.817300 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.827512 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.827836 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dvfr4" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="registry-server" containerID="cri-o://3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d" gracePeriod=30 Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.829382 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlzlc"] Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.869951 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.870443 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.870490 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p78sw\" (UniqueName: \"kubernetes.io/projected/7fa8126c-8308-4ce3-8529-70b28a606428-kube-api-access-p78sw\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.972016 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.972108 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.972158 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p78sw\" (UniqueName: \"kubernetes.io/projected/7fa8126c-8308-4ce3-8529-70b28a606428-kube-api-access-p78sw\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.973809 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.978349 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/7fa8126c-8308-4ce3-8529-70b28a606428-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:12 crc kubenswrapper[4708]: I1125 05:44:12.986962 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p78sw\" (UniqueName: \"kubernetes.io/projected/7fa8126c-8308-4ce3-8529-70b28a606428-kube-api-access-p78sw\") pod \"marketplace-operator-79b997595-qlzlc\" (UID: \"7fa8126c-8308-4ce3-8529-70b28a606428\") " pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.077548 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.173161 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.176256 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content\") pod \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.176377 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities\") pod \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.176536 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v95j7\" (UniqueName: \"kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7\") pod \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\" (UID: \"03bf86cb-ba81-4f29-92b8-8a3456e5d822\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.177499 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities" (OuterVolumeSpecName: "utilities") pod "03bf86cb-ba81-4f29-92b8-8a3456e5d822" (UID: "03bf86cb-ba81-4f29-92b8-8a3456e5d822"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.184850 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7" (OuterVolumeSpecName: "kube-api-access-v95j7") pod "03bf86cb-ba81-4f29-92b8-8a3456e5d822" (UID: "03bf86cb-ba81-4f29-92b8-8a3456e5d822"). InnerVolumeSpecName "kube-api-access-v95j7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.202502 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.227746 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03bf86cb-ba81-4f29-92b8-8a3456e5d822" (UID: "03bf86cb-ba81-4f29-92b8-8a3456e5d822"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.244688 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.253368 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277768 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics\") pod \"115b7e0c-95db-4ead-8f17-d1b9941bc933\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277821 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities\") pod \"22b89809-0b22-4ec3-8024-afd12a5bbb64\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277855 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca\") pod \"115b7e0c-95db-4ead-8f17-d1b9941bc933\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277888 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content\") pod \"22b89809-0b22-4ec3-8024-afd12a5bbb64\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277914 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content\") pod \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277950 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pk9s\" (UniqueName: \"kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s\") pod \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.277979 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jtk2r\" (UniqueName: \"kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r\") pod \"22b89809-0b22-4ec3-8024-afd12a5bbb64\" (UID: \"22b89809-0b22-4ec3-8024-afd12a5bbb64\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.278028 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities\") pod \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\" (UID: \"c69c6dd4-4e06-4343-add6-a78ebd99ac49\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.278053 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gngcd\" (UniqueName: \"kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd\") pod \"115b7e0c-95db-4ead-8f17-d1b9941bc933\" (UID: \"115b7e0c-95db-4ead-8f17-d1b9941bc933\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.278196 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v95j7\" (UniqueName: \"kubernetes.io/projected/03bf86cb-ba81-4f29-92b8-8a3456e5d822-kube-api-access-v95j7\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.278222 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.278233 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03bf86cb-ba81-4f29-92b8-8a3456e5d822-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.279756 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities" (OuterVolumeSpecName: "utilities") pod "22b89809-0b22-4ec3-8024-afd12a5bbb64" (UID: "22b89809-0b22-4ec3-8024-afd12a5bbb64"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.282671 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities" (OuterVolumeSpecName: "utilities") pod "c69c6dd4-4e06-4343-add6-a78ebd99ac49" (UID: "c69c6dd4-4e06-4343-add6-a78ebd99ac49"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.282695 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "115b7e0c-95db-4ead-8f17-d1b9941bc933" (UID: "115b7e0c-95db-4ead-8f17-d1b9941bc933"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.286322 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "115b7e0c-95db-4ead-8f17-d1b9941bc933" (UID: "115b7e0c-95db-4ead-8f17-d1b9941bc933"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.287705 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd" (OuterVolumeSpecName: "kube-api-access-gngcd") pod "115b7e0c-95db-4ead-8f17-d1b9941bc933" (UID: "115b7e0c-95db-4ead-8f17-d1b9941bc933"). InnerVolumeSpecName "kube-api-access-gngcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.287800 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r" (OuterVolumeSpecName: "kube-api-access-jtk2r") pod "22b89809-0b22-4ec3-8024-afd12a5bbb64" (UID: "22b89809-0b22-4ec3-8024-afd12a5bbb64"). InnerVolumeSpecName "kube-api-access-jtk2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.287990 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s" (OuterVolumeSpecName: "kube-api-access-6pk9s") pod "c69c6dd4-4e06-4343-add6-a78ebd99ac49" (UID: "c69c6dd4-4e06-4343-add6-a78ebd99ac49"). InnerVolumeSpecName "kube-api-access-6pk9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.301029 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22b89809-0b22-4ec3-8024-afd12a5bbb64" (UID: "22b89809-0b22-4ec3-8024-afd12a5bbb64"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.361586 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c69c6dd4-4e06-4343-add6-a78ebd99ac49" (UID: "c69c6dd4-4e06-4343-add6-a78ebd99ac49"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.378719 4708 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.378812 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.378891 4708 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/115b7e0c-95db-4ead-8f17-d1b9941bc933-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.378978 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22b89809-0b22-4ec3-8024-afd12a5bbb64-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.379072 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.379148 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pk9s\" (UniqueName: \"kubernetes.io/projected/c69c6dd4-4e06-4343-add6-a78ebd99ac49-kube-api-access-6pk9s\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.379224 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jtk2r\" (UniqueName: \"kubernetes.io/projected/22b89809-0b22-4ec3-8024-afd12a5bbb64-kube-api-access-jtk2r\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.379298 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c69c6dd4-4e06-4343-add6-a78ebd99ac49-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.379378 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gngcd\" (UniqueName: \"kubernetes.io/projected/115b7e0c-95db-4ead-8f17-d1b9941bc933-kube-api-access-gngcd\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.394979 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.479605 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities\") pod \"f1cb9071-9924-4611-b46b-6783b59eb4cd\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.479770 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5n5gk\" (UniqueName: \"kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk\") pod \"f1cb9071-9924-4611-b46b-6783b59eb4cd\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.479857 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content\") pod \"f1cb9071-9924-4611-b46b-6783b59eb4cd\" (UID: \"f1cb9071-9924-4611-b46b-6783b59eb4cd\") " Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.480138 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities" (OuterVolumeSpecName: "utilities") pod "f1cb9071-9924-4611-b46b-6783b59eb4cd" (UID: "f1cb9071-9924-4611-b46b-6783b59eb4cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.483283 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk" (OuterVolumeSpecName: "kube-api-access-5n5gk") pod "f1cb9071-9924-4611-b46b-6783b59eb4cd" (UID: "f1cb9071-9924-4611-b46b-6783b59eb4cd"). InnerVolumeSpecName "kube-api-access-5n5gk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.517567 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1cb9071-9924-4611-b46b-6783b59eb4cd" (UID: "f1cb9071-9924-4611-b46b-6783b59eb4cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.526102 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qlzlc"] Nov 25 05:44:13 crc kubenswrapper[4708]: W1125 05:44:13.529166 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fa8126c_8308_4ce3_8529_70b28a606428.slice/crio-829ad8e93fdce6d2decfa776fc713877b3262ac8483eb11bbe8f1ad921d2f990 WatchSource:0}: Error finding container 829ad8e93fdce6d2decfa776fc713877b3262ac8483eb11bbe8f1ad921d2f990: Status 404 returned error can't find the container with id 829ad8e93fdce6d2decfa776fc713877b3262ac8483eb11bbe8f1ad921d2f990 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.580906 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.580956 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5n5gk\" (UniqueName: \"kubernetes.io/projected/f1cb9071-9924-4611-b46b-6783b59eb4cd-kube-api-access-5n5gk\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.580968 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1cb9071-9924-4611-b46b-6783b59eb4cd-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.850938 4708 generic.go:334] "Generic (PLEG): container finished" podID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerID="3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d" exitCode=0 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.851037 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dvfr4" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.851093 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerDied","Data":"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.851214 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dvfr4" event={"ID":"c69c6dd4-4e06-4343-add6-a78ebd99ac49","Type":"ContainerDied","Data":"02eca66e4dc5b85de8794c08d8f3333cd57506e36e93879d53904f3bfbf300df"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.851244 4708 scope.go:117] "RemoveContainer" containerID="3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.854266 4708 generic.go:334] "Generic (PLEG): container finished" podID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerID="7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab" exitCode=0 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.854326 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.854343 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" event={"ID":"115b7e0c-95db-4ead-8f17-d1b9941bc933","Type":"ContainerDied","Data":"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.854377 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mr2tz" event={"ID":"115b7e0c-95db-4ead-8f17-d1b9941bc933","Type":"ContainerDied","Data":"44a48f8bfd99cc5c30c395ae25e09602aa7932e6295156c563effea38271ce96"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.856822 4708 generic.go:334] "Generic (PLEG): container finished" podID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerID="589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c" exitCode=0 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.856869 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerDied","Data":"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.856916 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6m7c" event={"ID":"22b89809-0b22-4ec3-8024-afd12a5bbb64","Type":"ContainerDied","Data":"f77c4ddc10e530ad824cdec284fc0dfa01df60238b9589a3e5b3c0d7c1f01a30"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.857084 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6m7c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.859206 4708 generic.go:334] "Generic (PLEG): container finished" podID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerID="2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e" exitCode=0 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.859265 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerDied","Data":"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.859290 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fz4gp" event={"ID":"f1cb9071-9924-4611-b46b-6783b59eb4cd","Type":"ContainerDied","Data":"0b9b70c34341ab06f17f7243ceb62198defeacafdec869f255146e703065320f"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.859358 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fz4gp" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.862200 4708 generic.go:334] "Generic (PLEG): container finished" podID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerID="249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484" exitCode=0 Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.862267 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerDied","Data":"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.862284 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m5q65" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.862296 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m5q65" event={"ID":"03bf86cb-ba81-4f29-92b8-8a3456e5d822","Type":"ContainerDied","Data":"4d82e5ead400db115aca12c59894e8df50079e9d4f225877ced4aabb39bbbdc2"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.864244 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" event={"ID":"7fa8126c-8308-4ce3-8529-70b28a606428","Type":"ContainerStarted","Data":"54dde1f1e5df47b461f9724f6dc8b4d24d6a933f3ba7de5657740c04aa2d7787"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.864278 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" event={"ID":"7fa8126c-8308-4ce3-8529-70b28a606428","Type":"ContainerStarted","Data":"829ad8e93fdce6d2decfa776fc713877b3262ac8483eb11bbe8f1ad921d2f990"} Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.864451 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.865535 4708 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qlzlc container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" start-of-body= Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.865602 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" podUID="7fa8126c-8308-4ce3-8529-70b28a606428" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.57:8080/healthz\": dial tcp 10.217.0.57:8080: connect: connection refused" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.873885 4708 scope.go:117] "RemoveContainer" containerID="0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.900327 4708 scope.go:117] "RemoveContainer" containerID="8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.901615 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" podStartSLOduration=1.901595254 podStartE2EDuration="1.901595254s" podCreationTimestamp="2025-11-25 05:44:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:44:13.899161618 +0000 UTC m=+195.307995004" watchObservedRunningTime="2025-11-25 05:44:13.901595254 +0000 UTC m=+195.310428640" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.907483 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.910885 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mr2tz"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.927945 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.930702 4708 scope.go:117] "RemoveContainer" containerID="3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d" Nov 25 05:44:13 crc kubenswrapper[4708]: E1125 05:44:13.933911 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d\": container with ID starting with 3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d not found: ID does not exist" containerID="3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.933964 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d"} err="failed to get container status \"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d\": rpc error: code = NotFound desc = could not find container \"3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d\": container with ID starting with 3203284f468c217dd448d0f50e17f700c69b093abdb4cd72b5171abe2530fd9d not found: ID does not exist" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.934178 4708 scope.go:117] "RemoveContainer" containerID="0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c" Nov 25 05:44:13 crc kubenswrapper[4708]: E1125 05:44:13.936963 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c\": container with ID starting with 0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c not found: ID does not exist" containerID="0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.936987 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c"} err="failed to get container status \"0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c\": rpc error: code = NotFound desc = could not find container \"0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c\": container with ID starting with 0abbd07ba1b5e4fcd0861630fdd408529685762eae24dce47ad29a384e09787c not found: ID does not exist" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.937002 4708 scope.go:117] "RemoveContainer" containerID="8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.937209 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dvfr4"] Nov 25 05:44:13 crc kubenswrapper[4708]: E1125 05:44:13.937303 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c\": container with ID starting with 8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c not found: ID does not exist" containerID="8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.937326 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c"} err="failed to get container status \"8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c\": rpc error: code = NotFound desc = could not find container \"8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c\": container with ID starting with 8e8340b26b6f2fbeffd59bf23e9ec0856c783f2dc4540f692bcabf8eaeb2061c not found: ID does not exist" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.937338 4708 scope.go:117] "RemoveContainer" containerID="7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.942056 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.947004 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6m7c"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.951599 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.953647 4708 scope.go:117] "RemoveContainer" containerID="7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.953756 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m5q65"] Nov 25 05:44:13 crc kubenswrapper[4708]: E1125 05:44:13.954051 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab\": container with ID starting with 7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab not found: ID does not exist" containerID="7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.954077 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab"} err="failed to get container status \"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab\": rpc error: code = NotFound desc = could not find container \"7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab\": container with ID starting with 7339db3dfdba882bef4a6e67e6946d9077362fe19cee35cdd4517d7a7c4c06ab not found: ID does not exist" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.954092 4708 scope.go:117] "RemoveContainer" containerID="589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.964967 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.965016 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fz4gp"] Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.968798 4708 scope.go:117] "RemoveContainer" containerID="48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe" Nov 25 05:44:13 crc kubenswrapper[4708]: I1125 05:44:13.990089 4708 scope.go:117] "RemoveContainer" containerID="93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.005204 4708 scope.go:117] "RemoveContainer" containerID="589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.006071 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c\": container with ID starting with 589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c not found: ID does not exist" containerID="589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.006136 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c"} err="failed to get container status \"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c\": rpc error: code = NotFound desc = could not find container \"589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c\": container with ID starting with 589ed71d1f1b8b73c76195e2c96ba2e8e7df68ef7cebdb302a86a61ae9abb03c not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.006363 4708 scope.go:117] "RemoveContainer" containerID="48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.006943 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe\": container with ID starting with 48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe not found: ID does not exist" containerID="48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.007070 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe"} err="failed to get container status \"48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe\": rpc error: code = NotFound desc = could not find container \"48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe\": container with ID starting with 48c17f9e4dd7291ba9346ea6bf655e72a538da8bcdab5afd14c6dec9be9d7dfe not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.007157 4708 scope.go:117] "RemoveContainer" containerID="93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.007627 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a\": container with ID starting with 93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a not found: ID does not exist" containerID="93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.007656 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a"} err="failed to get container status \"93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a\": rpc error: code = NotFound desc = could not find container \"93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a\": container with ID starting with 93c6136fea0a4678c4496531514e5663c0a5bd7e5a02ff1007a39c0c3ae31c2a not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.007676 4708 scope.go:117] "RemoveContainer" containerID="2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.020271 4708 scope.go:117] "RemoveContainer" containerID="6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.035176 4708 scope.go:117] "RemoveContainer" containerID="5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.046102 4708 scope.go:117] "RemoveContainer" containerID="2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.046357 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e\": container with ID starting with 2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e not found: ID does not exist" containerID="2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.046451 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e"} err="failed to get container status \"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e\": rpc error: code = NotFound desc = could not find container \"2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e\": container with ID starting with 2267b035d94f0bf29f975ee76e54c07ca10aa4b1f99b2683a54c1893d70a0b9e not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.046570 4708 scope.go:117] "RemoveContainer" containerID="6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.046842 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a\": container with ID starting with 6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a not found: ID does not exist" containerID="6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.046948 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a"} err="failed to get container status \"6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a\": rpc error: code = NotFound desc = could not find container \"6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a\": container with ID starting with 6d6df34ffcbade251e1de3234c94a931c060e2c49fa560f949802c459d821b9a not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.047030 4708 scope.go:117] "RemoveContainer" containerID="5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.047304 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241\": container with ID starting with 5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241 not found: ID does not exist" containerID="5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.047399 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241"} err="failed to get container status \"5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241\": rpc error: code = NotFound desc = could not find container \"5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241\": container with ID starting with 5b5a6732b7857745dfe3f4f66087d4929195829b96073fd61949742b97089241 not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.047468 4708 scope.go:117] "RemoveContainer" containerID="249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.059053 4708 scope.go:117] "RemoveContainer" containerID="05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.073359 4708 scope.go:117] "RemoveContainer" containerID="1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.086325 4708 scope.go:117] "RemoveContainer" containerID="249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.086977 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484\": container with ID starting with 249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484 not found: ID does not exist" containerID="249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.087040 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484"} err="failed to get container status \"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484\": rpc error: code = NotFound desc = could not find container \"249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484\": container with ID starting with 249ab979a00fca6d5f30a8ed4d08e652c5a5c73c44d398ec42400ea644166484 not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.087061 4708 scope.go:117] "RemoveContainer" containerID="05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.088176 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb\": container with ID starting with 05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb not found: ID does not exist" containerID="05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.088493 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb"} err="failed to get container status \"05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb\": rpc error: code = NotFound desc = could not find container \"05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb\": container with ID starting with 05269d85c345159928fe3862c87d45f7d6cc5e9779ceddd5747bef0940f3ffdb not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.088542 4708 scope.go:117] "RemoveContainer" containerID="1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.088911 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95\": container with ID starting with 1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95 not found: ID does not exist" containerID="1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.088951 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95"} err="failed to get container status \"1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95\": rpc error: code = NotFound desc = could not find container \"1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95\": container with ID starting with 1f0fd135220548a2f6a64320a30be1b46956e8a3bb9b3c26c6fba6348d41fd95 not found: ID does not exist" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.357354 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.357425 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.884356 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qlzlc" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.898559 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" path="/var/lib/kubelet/pods/03bf86cb-ba81-4f29-92b8-8a3456e5d822/volumes" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.899267 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" path="/var/lib/kubelet/pods/115b7e0c-95db-4ead-8f17-d1b9941bc933/volumes" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.899743 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" path="/var/lib/kubelet/pods/22b89809-0b22-4ec3-8024-afd12a5bbb64/volumes" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.900302 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" path="/var/lib/kubelet/pods/c69c6dd4-4e06-4343-add6-a78ebd99ac49/volumes" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.901572 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" path="/var/lib/kubelet/pods/f1cb9071-9924-4611-b46b-6783b59eb4cd/volumes" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986450 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8fwz"] Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986671 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986685 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986695 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986701 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986711 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986717 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986725 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986730 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986739 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986746 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986752 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986758 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986764 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986769 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986777 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986783 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986791 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986797 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986804 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986812 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="extract-utilities" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986818 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986823 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986831 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986836 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="extract-content" Nov 25 05:44:14 crc kubenswrapper[4708]: E1125 05:44:14.986845 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerName="marketplace-operator" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986850 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerName="marketplace-operator" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986945 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="03bf86cb-ba81-4f29-92b8-8a3456e5d822" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986953 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="115b7e0c-95db-4ead-8f17-d1b9941bc933" containerName="marketplace-operator" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986960 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1cb9071-9924-4611-b46b-6783b59eb4cd" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986971 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="22b89809-0b22-4ec3-8024-afd12a5bbb64" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.986982 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c69c6dd4-4e06-4343-add6-a78ebd99ac49" containerName="registry-server" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.987645 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:14 crc kubenswrapper[4708]: I1125 05:44:14.990323 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:14.999607 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8fwz"] Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.101401 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grcf2\" (UniqueName: \"kubernetes.io/projected/4295a1d6-dc34-4671-be69-a6c0a6b5f698-kube-api-access-grcf2\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.101446 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-utilities\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.101493 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-catalog-content\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.193848 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w9wdd"] Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.194898 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.196617 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.200549 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w9wdd"] Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.203014 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-utilities\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.203117 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-catalog-content\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.203201 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grcf2\" (UniqueName: \"kubernetes.io/projected/4295a1d6-dc34-4671-be69-a6c0a6b5f698-kube-api-access-grcf2\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.203478 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-utilities\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.203577 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4295a1d6-dc34-4671-be69-a6c0a6b5f698-catalog-content\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.221612 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grcf2\" (UniqueName: \"kubernetes.io/projected/4295a1d6-dc34-4671-be69-a6c0a6b5f698-kube-api-access-grcf2\") pod \"redhat-marketplace-b8fwz\" (UID: \"4295a1d6-dc34-4671-be69-a6c0a6b5f698\") " pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.303547 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.304003 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-utilities\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.304077 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-catalog-content\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.304132 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhqrd\" (UniqueName: \"kubernetes.io/projected/8bf634fc-fea4-4334-9015-38cb2e07095a-kube-api-access-hhqrd\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.406091 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-catalog-content\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.406413 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhqrd\" (UniqueName: \"kubernetes.io/projected/8bf634fc-fea4-4334-9015-38cb2e07095a-kube-api-access-hhqrd\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.406585 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-utilities\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.406609 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-catalog-content\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.406979 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bf634fc-fea4-4334-9015-38cb2e07095a-utilities\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.423229 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhqrd\" (UniqueName: \"kubernetes.io/projected/8bf634fc-fea4-4334-9015-38cb2e07095a-kube-api-access-hhqrd\") pod \"redhat-operators-w9wdd\" (UID: \"8bf634fc-fea4-4334-9015-38cb2e07095a\") " pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.510582 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.691035 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8fwz"] Nov 25 05:44:15 crc kubenswrapper[4708]: W1125 05:44:15.697177 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4295a1d6_dc34_4671_be69_a6c0a6b5f698.slice/crio-5df3248ff57bdd352b52f6298e40dcbb1c087faf62a692ac6ed67fd649876fa4 WatchSource:0}: Error finding container 5df3248ff57bdd352b52f6298e40dcbb1c087faf62a692ac6ed67fd649876fa4: Status 404 returned error can't find the container with id 5df3248ff57bdd352b52f6298e40dcbb1c087faf62a692ac6ed67fd649876fa4 Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.856444 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w9wdd"] Nov 25 05:44:15 crc kubenswrapper[4708]: W1125 05:44:15.871235 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bf634fc_fea4_4334_9015_38cb2e07095a.slice/crio-22e7cfdd3a90d6e75498c9446959dcd64153dc8a44f693ed158e8ed50797cdf2 WatchSource:0}: Error finding container 22e7cfdd3a90d6e75498c9446959dcd64153dc8a44f693ed158e8ed50797cdf2: Status 404 returned error can't find the container with id 22e7cfdd3a90d6e75498c9446959dcd64153dc8a44f693ed158e8ed50797cdf2 Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.887339 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9wdd" event={"ID":"8bf634fc-fea4-4334-9015-38cb2e07095a","Type":"ContainerStarted","Data":"22e7cfdd3a90d6e75498c9446959dcd64153dc8a44f693ed158e8ed50797cdf2"} Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.889797 4708 generic.go:334] "Generic (PLEG): container finished" podID="4295a1d6-dc34-4671-be69-a6c0a6b5f698" containerID="c5575ee74d1bda249a187a9367873fbb2df3a02d0e5686938299bc3d22b11ba3" exitCode=0 Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.890364 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8fwz" event={"ID":"4295a1d6-dc34-4671-be69-a6c0a6b5f698","Type":"ContainerDied","Data":"c5575ee74d1bda249a187a9367873fbb2df3a02d0e5686938299bc3d22b11ba3"} Nov 25 05:44:15 crc kubenswrapper[4708]: I1125 05:44:15.890396 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8fwz" event={"ID":"4295a1d6-dc34-4671-be69-a6c0a6b5f698","Type":"ContainerStarted","Data":"5df3248ff57bdd352b52f6298e40dcbb1c087faf62a692ac6ed67fd649876fa4"} Nov 25 05:44:16 crc kubenswrapper[4708]: I1125 05:44:16.895587 4708 generic.go:334] "Generic (PLEG): container finished" podID="4295a1d6-dc34-4671-be69-a6c0a6b5f698" containerID="09017db780df1b56be38c52e4acb9b368147a127bb7ca08da0757096470ac076" exitCode=0 Nov 25 05:44:16 crc kubenswrapper[4708]: I1125 05:44:16.897363 4708 generic.go:334] "Generic (PLEG): container finished" podID="8bf634fc-fea4-4334-9015-38cb2e07095a" containerID="a3cfea56d10a67991dc042d06ba5127a089a4167884415565e9fc4d0dbdd81f8" exitCode=0 Nov 25 05:44:16 crc kubenswrapper[4708]: I1125 05:44:16.897542 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8fwz" event={"ID":"4295a1d6-dc34-4671-be69-a6c0a6b5f698","Type":"ContainerDied","Data":"09017db780df1b56be38c52e4acb9b368147a127bb7ca08da0757096470ac076"} Nov 25 05:44:16 crc kubenswrapper[4708]: I1125 05:44:16.897587 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9wdd" event={"ID":"8bf634fc-fea4-4334-9015-38cb2e07095a","Type":"ContainerDied","Data":"a3cfea56d10a67991dc042d06ba5127a089a4167884415565e9fc4d0dbdd81f8"} Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.389347 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9w72p"] Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.393175 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.394846 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.396556 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9w72p"] Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.431319 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-utilities\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.431394 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkslv\" (UniqueName: \"kubernetes.io/projected/3aa64bad-9407-4512-ac3c-ab4c0d27a896-kube-api-access-dkslv\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.431427 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-catalog-content\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.532724 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-catalog-content\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.532843 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-utilities\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.532877 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkslv\" (UniqueName: \"kubernetes.io/projected/3aa64bad-9407-4512-ac3c-ab4c0d27a896-kube-api-access-dkslv\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.533120 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-catalog-content\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.533401 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3aa64bad-9407-4512-ac3c-ab4c0d27a896-utilities\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.550604 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkslv\" (UniqueName: \"kubernetes.io/projected/3aa64bad-9407-4512-ac3c-ab4c0d27a896-kube-api-access-dkslv\") pod \"community-operators-9w72p\" (UID: \"3aa64bad-9407-4512-ac3c-ab4c0d27a896\") " pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.589730 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6ffgp"] Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.594039 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.594921 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ffgp"] Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.595379 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.737341 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-utilities\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.737372 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.737409 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-catalog-content\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.737444 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24n9v\" (UniqueName: \"kubernetes.io/projected/e85ce5fc-a263-4980-8186-4741cb651bbf-kube-api-access-24n9v\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.838618 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-utilities\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.838806 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-catalog-content\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.838832 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24n9v\" (UniqueName: \"kubernetes.io/projected/e85ce5fc-a263-4980-8186-4741cb651bbf-kube-api-access-24n9v\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.839063 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-utilities\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.839231 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e85ce5fc-a263-4980-8186-4741cb651bbf-catalog-content\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.856624 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24n9v\" (UniqueName: \"kubernetes.io/projected/e85ce5fc-a263-4980-8186-4741cb651bbf-kube-api-access-24n9v\") pod \"certified-operators-6ffgp\" (UID: \"e85ce5fc-a263-4980-8186-4741cb651bbf\") " pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.904016 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9wdd" event={"ID":"8bf634fc-fea4-4334-9015-38cb2e07095a","Type":"ContainerStarted","Data":"cb2f76c997a896916e008bf8b628deaf8c6cc9de58ef9d76d8b75d0ab66dd51f"} Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.908114 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8fwz" event={"ID":"4295a1d6-dc34-4671-be69-a6c0a6b5f698","Type":"ContainerStarted","Data":"37571a9a62ace581f01f7939c67219afbe483ec5bda05f32f868a4771e61e2ee"} Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.912931 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.925008 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9w72p"] Nov 25 05:44:17 crc kubenswrapper[4708]: I1125 05:44:17.938107 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8fwz" podStartSLOduration=2.453066058 podStartE2EDuration="3.938080972s" podCreationTimestamp="2025-11-25 05:44:14 +0000 UTC" firstStartedPulling="2025-11-25 05:44:15.890864346 +0000 UTC m=+197.299697732" lastFinishedPulling="2025-11-25 05:44:17.37587926 +0000 UTC m=+198.784712646" observedRunningTime="2025-11-25 05:44:17.935275685 +0000 UTC m=+199.344109070" watchObservedRunningTime="2025-11-25 05:44:17.938080972 +0000 UTC m=+199.346914357" Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.101268 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ffgp"] Nov 25 05:44:18 crc kubenswrapper[4708]: W1125 05:44:18.134900 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode85ce5fc_a263_4980_8186_4741cb651bbf.slice/crio-0fcb8a3b2fadd7b5356aafc27cd6862805b7147b18e295b71c82536d324ef3a4 WatchSource:0}: Error finding container 0fcb8a3b2fadd7b5356aafc27cd6862805b7147b18e295b71c82536d324ef3a4: Status 404 returned error can't find the container with id 0fcb8a3b2fadd7b5356aafc27cd6862805b7147b18e295b71c82536d324ef3a4 Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.921791 4708 generic.go:334] "Generic (PLEG): container finished" podID="e85ce5fc-a263-4980-8186-4741cb651bbf" containerID="c4c6582004ea76f8f70a60dcf4223a7cbfb5349259595c14eba976489be27238" exitCode=0 Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.921891 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ffgp" event={"ID":"e85ce5fc-a263-4980-8186-4741cb651bbf","Type":"ContainerDied","Data":"c4c6582004ea76f8f70a60dcf4223a7cbfb5349259595c14eba976489be27238"} Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.922184 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ffgp" event={"ID":"e85ce5fc-a263-4980-8186-4741cb651bbf","Type":"ContainerStarted","Data":"0fcb8a3b2fadd7b5356aafc27cd6862805b7147b18e295b71c82536d324ef3a4"} Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.923637 4708 generic.go:334] "Generic (PLEG): container finished" podID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" containerID="c73a4e14a12a329734cdc3114b359be8a82c0942905b6465e133b45f1d016696" exitCode=0 Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.923718 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w72p" event={"ID":"3aa64bad-9407-4512-ac3c-ab4c0d27a896","Type":"ContainerDied","Data":"c73a4e14a12a329734cdc3114b359be8a82c0942905b6465e133b45f1d016696"} Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.923738 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w72p" event={"ID":"3aa64bad-9407-4512-ac3c-ab4c0d27a896","Type":"ContainerStarted","Data":"b88db5cf0cf831395764f56140a1c81d5199527eabcbaf5dea3ee50e1c04c8dd"} Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.925454 4708 generic.go:334] "Generic (PLEG): container finished" podID="8bf634fc-fea4-4334-9015-38cb2e07095a" containerID="cb2f76c997a896916e008bf8b628deaf8c6cc9de58ef9d76d8b75d0ab66dd51f" exitCode=0 Nov 25 05:44:18 crc kubenswrapper[4708]: I1125 05:44:18.925620 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9wdd" event={"ID":"8bf634fc-fea4-4334-9015-38cb2e07095a","Type":"ContainerDied","Data":"cb2f76c997a896916e008bf8b628deaf8c6cc9de58ef9d76d8b75d0ab66dd51f"} Nov 25 05:44:19 crc kubenswrapper[4708]: I1125 05:44:19.931807 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9wdd" event={"ID":"8bf634fc-fea4-4334-9015-38cb2e07095a","Type":"ContainerStarted","Data":"7adaa96808041a710f9993fa71954ebaed021a4ab880ff44f8dbf42f73264214"} Nov 25 05:44:19 crc kubenswrapper[4708]: I1125 05:44:19.950352 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w9wdd" podStartSLOduration=2.485544333 podStartE2EDuration="4.95032642s" podCreationTimestamp="2025-11-25 05:44:15 +0000 UTC" firstStartedPulling="2025-11-25 05:44:16.898259246 +0000 UTC m=+198.307092632" lastFinishedPulling="2025-11-25 05:44:19.363041333 +0000 UTC m=+200.771874719" observedRunningTime="2025-11-25 05:44:19.945817231 +0000 UTC m=+201.354650617" watchObservedRunningTime="2025-11-25 05:44:19.95032642 +0000 UTC m=+201.359159805" Nov 25 05:44:20 crc kubenswrapper[4708]: I1125 05:44:20.938447 4708 generic.go:334] "Generic (PLEG): container finished" podID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" containerID="e8df490a8d10c6a5b3abe50ab9d4dc3f80e8d18e2578518737f59d21eaa52a61" exitCode=0 Nov 25 05:44:20 crc kubenswrapper[4708]: I1125 05:44:20.938562 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w72p" event={"ID":"3aa64bad-9407-4512-ac3c-ab4c0d27a896","Type":"ContainerDied","Data":"e8df490a8d10c6a5b3abe50ab9d4dc3f80e8d18e2578518737f59d21eaa52a61"} Nov 25 05:44:20 crc kubenswrapper[4708]: I1125 05:44:20.941596 4708 generic.go:334] "Generic (PLEG): container finished" podID="e85ce5fc-a263-4980-8186-4741cb651bbf" containerID="fce375a07a5fb334a969317ca4e417c1479f25626b684f825f4c32358800ed00" exitCode=0 Nov 25 05:44:20 crc kubenswrapper[4708]: I1125 05:44:20.941678 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ffgp" event={"ID":"e85ce5fc-a263-4980-8186-4741cb651bbf","Type":"ContainerDied","Data":"fce375a07a5fb334a969317ca4e417c1479f25626b684f825f4c32358800ed00"} Nov 25 05:44:21 crc kubenswrapper[4708]: I1125 05:44:21.960951 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9w72p" event={"ID":"3aa64bad-9407-4512-ac3c-ab4c0d27a896","Type":"ContainerStarted","Data":"39a05b8f17d7aec99702af924b57dc283184e9155380550dd8a5abb56de6c8da"} Nov 25 05:44:21 crc kubenswrapper[4708]: I1125 05:44:21.963122 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ffgp" event={"ID":"e85ce5fc-a263-4980-8186-4741cb651bbf","Type":"ContainerStarted","Data":"39f2ee99e28afe86f623588dfca7e6ab331e876318f697c8a48305d81f341292"} Nov 25 05:44:21 crc kubenswrapper[4708]: I1125 05:44:21.974379 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9w72p" podStartSLOduration=2.503550737 podStartE2EDuration="4.974363219s" podCreationTimestamp="2025-11-25 05:44:17 +0000 UTC" firstStartedPulling="2025-11-25 05:44:18.927929104 +0000 UTC m=+200.336762489" lastFinishedPulling="2025-11-25 05:44:21.398741585 +0000 UTC m=+202.807574971" observedRunningTime="2025-11-25 05:44:21.97279828 +0000 UTC m=+203.381631666" watchObservedRunningTime="2025-11-25 05:44:21.974363219 +0000 UTC m=+203.383196605" Nov 25 05:44:21 crc kubenswrapper[4708]: I1125 05:44:21.988058 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6ffgp" podStartSLOduration=2.49109932 podStartE2EDuration="4.988041087s" podCreationTimestamp="2025-11-25 05:44:17 +0000 UTC" firstStartedPulling="2025-11-25 05:44:18.923721162 +0000 UTC m=+200.332554539" lastFinishedPulling="2025-11-25 05:44:21.42066292 +0000 UTC m=+202.829496306" observedRunningTime="2025-11-25 05:44:21.983570001 +0000 UTC m=+203.392403386" watchObservedRunningTime="2025-11-25 05:44:21.988041087 +0000 UTC m=+203.396874473" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.303753 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.304745 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.338436 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.511675 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.511718 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:25 crc kubenswrapper[4708]: I1125 05:44:25.542557 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:26 crc kubenswrapper[4708]: I1125 05:44:26.009717 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8fwz" Nov 25 05:44:26 crc kubenswrapper[4708]: I1125 05:44:26.015079 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w9wdd" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.202428 4708 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203456 4708 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203623 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203757 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342" gracePeriod=15 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203793 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653" gracePeriod=15 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203821 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691" gracePeriod=15 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203801 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b" gracePeriod=15 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.203909 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02" gracePeriod=15 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.204779 4708 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205025 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205048 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205058 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205063 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205092 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205099 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205112 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205118 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205130 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205135 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205144 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205166 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 05:44:27 crc kubenswrapper[4708]: E1125 05:44:27.205174 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205180 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205329 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205353 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205361 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205368 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205378 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.205386 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.360903 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361189 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361248 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361283 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361457 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361550 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361612 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.361638 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462701 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462763 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462791 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462832 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462830 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.462855 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463107 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463130 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463143 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463107 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463158 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463182 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463210 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463240 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463255 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.463275 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.738289 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.738347 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.773918 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.774887 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.775281 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.913391 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.913425 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.944658 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.945190 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.945489 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.945753 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.990791 4708 generic.go:334] "Generic (PLEG): container finished" podID="7f521c0f-a241-4990-b984-ac475119e8d8" containerID="7c01b3977b7679972a415340e75baaae1074324090e5d0abe8ce38816c0efbea" exitCode=0 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.990872 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7f521c0f-a241-4990-b984-ac475119e8d8","Type":"ContainerDied","Data":"7c01b3977b7679972a415340e75baaae1074324090e5d0abe8ce38816c0efbea"} Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.991416 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.991680 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.991977 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.992292 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.994096 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995221 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995835 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653" exitCode=0 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995853 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691" exitCode=0 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995860 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b" exitCode=0 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995868 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02" exitCode=2 Nov 25 05:44:27 crc kubenswrapper[4708]: I1125 05:44:27.995940 4708 scope.go:117] "RemoveContainer" containerID="713f2bb5c3750040ad9d5be98408d3c0e24b970abcfb078f1c6e5ac9a5e50a9d" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.025735 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9w72p" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.026170 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.026397 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.026634 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.026885 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.027830 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6ffgp" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.028197 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.028388 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.028622 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.028873 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.130225 4708 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.130268 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.896769 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.897113 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.897469 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:28 crc kubenswrapper[4708]: I1125 05:44:28.897959 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.001624 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.245309 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.246189 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.246587 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.247142 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385120 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir\") pod \"7f521c0f-a241-4990-b984-ac475119e8d8\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385450 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock\") pod \"7f521c0f-a241-4990-b984-ac475119e8d8\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385507 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access\") pod \"7f521c0f-a241-4990-b984-ac475119e8d8\" (UID: \"7f521c0f-a241-4990-b984-ac475119e8d8\") " Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385206 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7f521c0f-a241-4990-b984-ac475119e8d8" (UID: "7f521c0f-a241-4990-b984-ac475119e8d8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385499 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock" (OuterVolumeSpecName: "var-lock") pod "7f521c0f-a241-4990-b984-ac475119e8d8" (UID: "7f521c0f-a241-4990-b984-ac475119e8d8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385756 4708 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.385773 4708 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/7f521c0f-a241-4990-b984-ac475119e8d8-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.391546 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7f521c0f-a241-4990-b984-ac475119e8d8" (UID: "7f521c0f-a241-4990-b984-ac475119e8d8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:44:29 crc kubenswrapper[4708]: I1125 05:44:29.486963 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7f521c0f-a241-4990-b984-ac475119e8d8-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.013160 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.013761 4708 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342" exitCode=0 Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.014854 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"7f521c0f-a241-4990-b984-ac475119e8d8","Type":"ContainerDied","Data":"f61d7477f4eab755ef27109a4dfdabc7689e75117e2f414255aca6d1b24113f5"} Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.014889 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f61d7477f4eab755ef27109a4dfdabc7689e75117e2f414255aca6d1b24113f5" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.014942 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.026015 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.026320 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.026496 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.052299 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.052855 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.053180 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.053444 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.053676 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.053861 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195064 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195124 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195163 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195184 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195230 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195318 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195447 4708 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195461 4708 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.195470 4708 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:44:30 crc kubenswrapper[4708]: I1125 05:44:30.898865 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.021622 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.022292 4708 scope.go:117] "RemoveContainer" containerID="fad2bdf0f67115c2f3c21d034a0d169061bb3b0fe203d6d5ca5684e0328be653" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.022346 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.023259 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.023581 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.023903 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.024233 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.024903 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.025134 4708 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.025438 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.025720 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.037781 4708 scope.go:117] "RemoveContainer" containerID="39cc9a82125f2de1c94c6aec562108b39d7b2dc7ca331d6d299500e43e7a1691" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.046883 4708 scope.go:117] "RemoveContainer" containerID="ad39d76413e2fb3387d9bfb86a0ace225f5b46a665162d8df143b579d207956b" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.054766 4708 scope.go:117] "RemoveContainer" containerID="a2f97b6d34cdd3a15c1a840ce5378961f47a56bcf5f7a3f13e74ea77e220ef02" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.066097 4708 scope.go:117] "RemoveContainer" containerID="bd410cdf63e13ebb8e74ab907f63459e9930d848abdc92f4b99d2a50c1cba342" Nov 25 05:44:31 crc kubenswrapper[4708]: I1125 05:44:31.082485 4708 scope.go:117] "RemoveContainer" containerID="c7575dc843cfbbdb45e19809f5218fca48fef2ef6b37ec7059e750dc3a4ba897" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.236813 4708 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.34:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:32 crc kubenswrapper[4708]: I1125 05:44:32.237146 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:32 crc kubenswrapper[4708]: W1125 05:44:32.256567 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-0b47cdf5f66ee679510f56d14e0a7cb50b996cbccb7db45fd0eab133c21a53b5 WatchSource:0}: Error finding container 0b47cdf5f66ee679510f56d14e0a7cb50b996cbccb7db45fd0eab133c21a53b5: Status 404 returned error can't find the container with id 0b47cdf5f66ee679510f56d14e0a7cb50b996cbccb7db45fd0eab133c21a53b5 Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.259238 4708 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 192.168.26.34:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b29a61031a1f3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 05:44:32.258679283 +0000 UTC m=+213.667512669,LastTimestamp:2025-11-25 05:44:32.258679283 +0000 UTC m=+213.667512669,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.750275 4708 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.750879 4708 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.751158 4708 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.751399 4708 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.751634 4708 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:32 crc kubenswrapper[4708]: I1125 05:44:32.751654 4708 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.751819 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="200ms" Nov 25 05:44:32 crc kubenswrapper[4708]: E1125 05:44:32.952588 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="400ms" Nov 25 05:44:33 crc kubenswrapper[4708]: I1125 05:44:33.039691 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb"} Nov 25 05:44:33 crc kubenswrapper[4708]: I1125 05:44:33.039740 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"0b47cdf5f66ee679510f56d14e0a7cb50b996cbccb7db45fd0eab133c21a53b5"} Nov 25 05:44:33 crc kubenswrapper[4708]: I1125 05:44:33.040319 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:33 crc kubenswrapper[4708]: E1125 05:44:33.040358 4708 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 192.168.26.34:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:44:33 crc kubenswrapper[4708]: I1125 05:44:33.040577 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:33 crc kubenswrapper[4708]: I1125 05:44:33.040811 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:33 crc kubenswrapper[4708]: E1125 05:44:33.353275 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="800ms" Nov 25 05:44:34 crc kubenswrapper[4708]: E1125 05:44:34.154594 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="1.6s" Nov 25 05:44:35 crc kubenswrapper[4708]: E1125 05:44:35.755461 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="3.2s" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.892708 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.894091 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.894375 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.894619 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.904416 4708 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.904441 4708 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:37 crc kubenswrapper[4708]: E1125 05:44:37.904707 4708 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:37 crc kubenswrapper[4708]: I1125 05:44:37.905123 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:37 crc kubenswrapper[4708]: W1125 05:44:37.920811 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-28cd2682f2fad9432e356d0465bd60c723e489b112ef8f7e2f70260d69635cc2 WatchSource:0}: Error finding container 28cd2682f2fad9432e356d0465bd60c723e489b112ef8f7e2f70260d69635cc2: Status 404 returned error can't find the container with id 28cd2682f2fad9432e356d0465bd60c723e489b112ef8f7e2f70260d69635cc2 Nov 25 05:44:38 crc kubenswrapper[4708]: I1125 05:44:38.069353 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"28cd2682f2fad9432e356d0465bd60c723e489b112ef8f7e2f70260d69635cc2"} Nov 25 05:44:38 crc kubenswrapper[4708]: I1125 05:44:38.898029 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:38 crc kubenswrapper[4708]: I1125 05:44:38.898367 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:38 crc kubenswrapper[4708]: I1125 05:44:38.898920 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:38 crc kubenswrapper[4708]: I1125 05:44:38.899167 4708 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:38 crc kubenswrapper[4708]: E1125 05:44:38.956164 4708 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 192.168.26.34:6443: connect: connection refused" interval="6.4s" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.086359 4708 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="d8b98957d9bb90bcd7692bc278357aecf3f31d2b63ccb85bdc13bd23ff37ea81" exitCode=0 Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.086411 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"d8b98957d9bb90bcd7692bc278357aecf3f31d2b63ccb85bdc13bd23ff37ea81"} Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.086728 4708 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.086747 4708 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:39 crc kubenswrapper[4708]: E1125 05:44:39.087178 4708 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.087565 4708 status_manager.go:851] "Failed to get status for pod" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.088166 4708 status_manager.go:851] "Failed to get status for pod" podUID="3aa64bad-9407-4512-ac3c-ab4c0d27a896" pod="openshift-marketplace/community-operators-9w72p" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/community-operators-9w72p\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.088463 4708 status_manager.go:851] "Failed to get status for pod" podUID="e85ce5fc-a263-4980-8186-4741cb651bbf" pod="openshift-marketplace/certified-operators-6ffgp" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/certified-operators-6ffgp\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.088720 4708 status_manager.go:851] "Failed to get status for pod" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 192.168.26.34:6443: connect: connection refused" Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.457582 4708 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Nov 25 05:44:39 crc kubenswrapper[4708]: I1125 05:44:39.457869 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.094506 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.094579 4708 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063" exitCode=1 Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.094640 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063"} Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.095201 4708 scope.go:117] "RemoveContainer" containerID="66906cbbfde607ccaaabd1acdf3b491383ff9241115ca388e3506758ed296063" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.098984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6c407b091a46fbc92df52959425c62735b9f799c3cc49e29eecf9ac90427c41d"} Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099126 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d3b6bccba9b55644e53f122afaf0c9057c3f608a90e26e0fb18b531f74d4cc9b"} Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099191 4708 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099212 4708 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099199 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099352 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3ec6a7641d818893ead84f68200726a4808cef9434f06fc8c9ee4bcae06aba79"} Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099413 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bfe156caea1d5483337e9a143b9be915012436231ab31d878bbecfd90c85529f"} Nov 25 05:44:40 crc kubenswrapper[4708]: I1125 05:44:40.099475 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"601d52f4174d87da3d3ca9e167ab625f2ddc013a3b5a0939aae4cf90ecd07833"} Nov 25 05:44:41 crc kubenswrapper[4708]: I1125 05:44:41.108736 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 05:44:41 crc kubenswrapper[4708]: I1125 05:44:41.109044 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"769b90126e69cbc768c304eb7601c456408cef6e8eb46f3e657b2bc410859f28"} Nov 25 05:44:42 crc kubenswrapper[4708]: I1125 05:44:42.905254 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:42 crc kubenswrapper[4708]: I1125 05:44:42.906036 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:42 crc kubenswrapper[4708]: I1125 05:44:42.910349 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:44 crc kubenswrapper[4708]: I1125 05:44:44.356863 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:44:44 crc kubenswrapper[4708]: I1125 05:44:44.357232 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:44:44 crc kubenswrapper[4708]: I1125 05:44:44.357281 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:44:44 crc kubenswrapper[4708]: I1125 05:44:44.357712 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:44:44 crc kubenswrapper[4708]: I1125 05:44:44.357764 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545" gracePeriod=600 Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.128746 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545" exitCode=0 Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.128817 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545"} Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.129090 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2"} Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.592410 4708 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.615436 4708 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5afeb74a-c361-4729-be2c-45bf284ef0db\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:44:39Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:44:39Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:44:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-cert-syncer kube-apiserver-cert-regeneration-controller kube-apiserver-insecure-readyz kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T05:44:37Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-cert-syncer kube-apiserver-cert-regeneration-controller kube-apiserver-insecure-readyz kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d8b98957d9bb90bcd7692bc278357aecf3f31d2b63ccb85bdc13bd23ff37ea81\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d8b98957d9bb90bcd7692bc278357aecf3f31d2b63ccb85bdc13bd23ff37ea81\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T05:44:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T05:44:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Pending\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Pod \"kube-apiserver-crc\" is invalid: metadata.uid: Invalid value: \"5afeb74a-c361-4729-be2c-45bf284ef0db\": field is immutable" Nov 25 05:44:45 crc kubenswrapper[4708]: I1125 05:44:45.694261 4708 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ee3c7e74-e793-4fbc-84f2-7591ee51c510" Nov 25 05:44:46 crc kubenswrapper[4708]: I1125 05:44:46.133381 4708 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:46 crc kubenswrapper[4708]: I1125 05:44:46.133413 4708 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:46 crc kubenswrapper[4708]: I1125 05:44:46.137314 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:44:46 crc kubenswrapper[4708]: I1125 05:44:46.137577 4708 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ee3c7e74-e793-4fbc-84f2-7591ee51c510" Nov 25 05:44:46 crc kubenswrapper[4708]: I1125 05:44:46.715478 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:44:47 crc kubenswrapper[4708]: I1125 05:44:47.137559 4708 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:47 crc kubenswrapper[4708]: I1125 05:44:47.137593 4708 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5afeb74a-c361-4729-be2c-45bf284ef0db" Nov 25 05:44:47 crc kubenswrapper[4708]: I1125 05:44:47.140958 4708 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ee3c7e74-e793-4fbc-84f2-7591ee51c510" Nov 25 05:44:48 crc kubenswrapper[4708]: I1125 05:44:48.281227 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:44:48 crc kubenswrapper[4708]: I1125 05:44:48.284656 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:44:52 crc kubenswrapper[4708]: I1125 05:44:52.730885 4708 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 05:44:53 crc kubenswrapper[4708]: I1125 05:44:53.269378 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 05:44:54 crc kubenswrapper[4708]: I1125 05:44:54.096511 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 05:44:54 crc kubenswrapper[4708]: I1125 05:44:54.594607 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 05:44:55 crc kubenswrapper[4708]: I1125 05:44:55.710584 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 05:44:55 crc kubenswrapper[4708]: I1125 05:44:55.712250 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 05:44:55 crc kubenswrapper[4708]: I1125 05:44:55.883919 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 05:44:56 crc kubenswrapper[4708]: I1125 05:44:56.429180 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 05:44:56 crc kubenswrapper[4708]: I1125 05:44:56.517510 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 05:44:56 crc kubenswrapper[4708]: I1125 05:44:56.544091 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 05:44:56 crc kubenswrapper[4708]: I1125 05:44:56.717781 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 05:44:57 crc kubenswrapper[4708]: I1125 05:44:57.088216 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 05:44:57 crc kubenswrapper[4708]: I1125 05:44:57.429440 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 05:44:57 crc kubenswrapper[4708]: I1125 05:44:57.694928 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 05:44:57 crc kubenswrapper[4708]: I1125 05:44:57.921629 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 05:44:57 crc kubenswrapper[4708]: I1125 05:44:57.990093 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 05:44:58 crc kubenswrapper[4708]: I1125 05:44:58.036857 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 05:44:58 crc kubenswrapper[4708]: I1125 05:44:58.074972 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 05:44:58 crc kubenswrapper[4708]: I1125 05:44:58.564311 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 05:44:58 crc kubenswrapper[4708]: I1125 05:44:58.720411 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 05:44:58 crc kubenswrapper[4708]: I1125 05:44:58.756546 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.007451 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.190201 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.416656 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.470190 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.616745 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 05:44:59 crc kubenswrapper[4708]: I1125 05:44:59.619955 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.044768 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.051237 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.418345 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.499764 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.611178 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.721320 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.805600 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.890367 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.921555 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 05:45:00 crc kubenswrapper[4708]: I1125 05:45:00.928967 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.198107 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.202865 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.229440 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.297433 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.366021 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.384688 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.453624 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.492898 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.493645 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.496989 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.590989 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.601002 4708 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.792271 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.861373 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.868380 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.872246 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.901985 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 05:45:01 crc kubenswrapper[4708]: I1125 05:45:01.906127 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.009107 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.022417 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.093453 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.121240 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.130272 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.135345 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.199795 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.253330 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.274792 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.353736 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.385327 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.409786 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.509219 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.543584 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.544604 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.576710 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.611583 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.663057 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.672505 4708 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.761691 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.778905 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 05:45:02 crc kubenswrapper[4708]: I1125 05:45:02.905120 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.011911 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.027931 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.102424 4708 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.132887 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.180947 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.239995 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.240111 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.259918 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.317398 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.320778 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.425133 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.471281 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.579607 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.624413 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.658464 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.728653 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.806993 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.824437 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.826331 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.857593 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.864757 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.935461 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.954462 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 05:45:03 crc kubenswrapper[4708]: I1125 05:45:03.997564 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.007816 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.061210 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.106221 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.146427 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.276840 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.297777 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.409509 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.477823 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.552085 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.587755 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.593325 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.597173 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.597308 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.925370 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.927542 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 05:45:04 crc kubenswrapper[4708]: I1125 05:45:04.995577 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.084273 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.092931 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.119136 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.215587 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.334131 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.401707 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.479057 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.479514 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.775198 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.783684 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 05:45:05 crc kubenswrapper[4708]: I1125 05:45:05.844733 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.077325 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.089267 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.191908 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.409539 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.417334 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.421431 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.483665 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.616415 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.627539 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.650377 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.722110 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.753070 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.845739 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.917916 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.944937 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 05:45:06 crc kubenswrapper[4708]: I1125 05:45:06.949219 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.040958 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.098537 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.106278 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.185685 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.205256 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.325982 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.383339 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.394169 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.456479 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.471407 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.549025 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.557035 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.562050 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.730648 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.754704 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.833515 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.843042 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 05:45:07 crc kubenswrapper[4708]: I1125 05:45:07.867220 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.087993 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.129343 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.224037 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.350581 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.409131 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.412229 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.436382 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.437707 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.473748 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.476947 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.515948 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.597510 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.653972 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.871069 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 05:45:08 crc kubenswrapper[4708]: I1125 05:45:08.891206 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.013744 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.054731 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.056360 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.144572 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.156115 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.177974 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.330132 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.357114 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.371548 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.459359 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.547390 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.573766 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.717324 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.753493 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.763893 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.787737 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.788025 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.810168 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 05:45:09 crc kubenswrapper[4708]: I1125 05:45:09.982107 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.017380 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.167717 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.168010 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.198734 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.247869 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.329641 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.334618 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.361948 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.384436 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.404655 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.491011 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.540514 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.566571 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.742985 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 05:45:10 crc kubenswrapper[4708]: I1125 05:45:10.904632 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.014752 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.129458 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.260926 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.298645 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.341094 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.411064 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.516338 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.657932 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.737882 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 05:45:11 crc kubenswrapper[4708]: I1125 05:45:11.874033 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.006310 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.054230 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.388684 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.424867 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.425101 4708 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.430375 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.430441 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2"] Nov 25 05:45:12 crc kubenswrapper[4708]: E1125 05:45:12.430720 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" containerName="installer" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.430740 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" containerName="installer" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.430874 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f521c0f-a241-4990-b984-ac475119e8d8" containerName="installer" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.431349 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.432700 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.432806 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.435877 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.448656 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=27.448617365 podStartE2EDuration="27.448617365s" podCreationTimestamp="2025-11-25 05:44:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:45:12.445863954 +0000 UTC m=+253.854697341" watchObservedRunningTime="2025-11-25 05:45:12.448617365 +0000 UTC m=+253.857450750" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.450210 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.597793 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptsf4\" (UniqueName: \"kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.597856 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.597898 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.601591 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.616962 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.617265 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.699149 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptsf4\" (UniqueName: \"kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.699824 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.699871 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.701415 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.706783 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.713180 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptsf4\" (UniqueName: \"kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4\") pod \"collect-profiles-29400825-kljt2\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.745414 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.952917 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.954591 4708 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 05:45:12 crc kubenswrapper[4708]: I1125 05:45:12.991257 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.065341 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.097308 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2"] Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.269659 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" event={"ID":"38ef401b-8325-4595-9d9e-dd6f0aca7078","Type":"ContainerStarted","Data":"f6f52a6a86f8fcb5169e062b0da706dca690087810b1ceaeb1bd339f5b983e42"} Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.270952 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" event={"ID":"38ef401b-8325-4595-9d9e-dd6f0aca7078","Type":"ContainerStarted","Data":"e89b0deed09efc3d7494519a694c68885560964d4affdda1d4745c2dc8a8dec5"} Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.283773 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" podStartSLOduration=10.283758866 podStartE2EDuration="10.283758866s" podCreationTimestamp="2025-11-25 05:45:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:45:13.281701057 +0000 UTC m=+254.690534444" watchObservedRunningTime="2025-11-25 05:45:13.283758866 +0000 UTC m=+254.692592252" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.314477 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.336478 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.439486 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.577699 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.760439 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.791810 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 05:45:13 crc kubenswrapper[4708]: I1125 05:45:13.846214 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.135056 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.157852 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.276108 4708 generic.go:334] "Generic (PLEG): container finished" podID="38ef401b-8325-4595-9d9e-dd6f0aca7078" containerID="f6f52a6a86f8fcb5169e062b0da706dca690087810b1ceaeb1bd339f5b983e42" exitCode=0 Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.276176 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" event={"ID":"38ef401b-8325-4595-9d9e-dd6f0aca7078","Type":"ContainerDied","Data":"f6f52a6a86f8fcb5169e062b0da706dca690087810b1ceaeb1bd339f5b983e42"} Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.374942 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.443829 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.452494 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.543362 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 05:45:14 crc kubenswrapper[4708]: I1125 05:45:14.869821 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.444328 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.482935 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.638480 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptsf4\" (UniqueName: \"kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4\") pod \"38ef401b-8325-4595-9d9e-dd6f0aca7078\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.638652 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume\") pod \"38ef401b-8325-4595-9d9e-dd6f0aca7078\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.638731 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume\") pod \"38ef401b-8325-4595-9d9e-dd6f0aca7078\" (UID: \"38ef401b-8325-4595-9d9e-dd6f0aca7078\") " Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.639294 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume" (OuterVolumeSpecName: "config-volume") pod "38ef401b-8325-4595-9d9e-dd6f0aca7078" (UID: "38ef401b-8325-4595-9d9e-dd6f0aca7078"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.644027 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "38ef401b-8325-4595-9d9e-dd6f0aca7078" (UID: "38ef401b-8325-4595-9d9e-dd6f0aca7078"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.644269 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4" (OuterVolumeSpecName: "kube-api-access-ptsf4") pod "38ef401b-8325-4595-9d9e-dd6f0aca7078" (UID: "38ef401b-8325-4595-9d9e-dd6f0aca7078"). InnerVolumeSpecName "kube-api-access-ptsf4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.681157 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.740023 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38ef401b-8325-4595-9d9e-dd6f0aca7078-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.740059 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptsf4\" (UniqueName: \"kubernetes.io/projected/38ef401b-8325-4595-9d9e-dd6f0aca7078-kube-api-access-ptsf4\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.740070 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/38ef401b-8325-4595-9d9e-dd6f0aca7078-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.823133 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 05:45:15 crc kubenswrapper[4708]: I1125 05:45:15.976568 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 05:45:16 crc kubenswrapper[4708]: I1125 05:45:16.288303 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" event={"ID":"38ef401b-8325-4595-9d9e-dd6f0aca7078","Type":"ContainerDied","Data":"e89b0deed09efc3d7494519a694c68885560964d4affdda1d4745c2dc8a8dec5"} Nov 25 05:45:16 crc kubenswrapper[4708]: I1125 05:45:16.288647 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e89b0deed09efc3d7494519a694c68885560964d4affdda1d4745c2dc8a8dec5" Nov 25 05:45:16 crc kubenswrapper[4708]: I1125 05:45:16.288349 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2" Nov 25 05:45:16 crc kubenswrapper[4708]: I1125 05:45:16.798636 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 05:45:18 crc kubenswrapper[4708]: I1125 05:45:18.336823 4708 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 05:45:18 crc kubenswrapper[4708]: I1125 05:45:18.337540 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb" gracePeriod=5 Nov 25 05:45:23 crc kubenswrapper[4708]: I1125 05:45:23.889698 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 05:45:23 crc kubenswrapper[4708]: I1125 05:45:23.889780 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030498 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030578 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030639 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030636 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030680 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030688 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030747 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030793 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.030897 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.031685 4708 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.031711 4708 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.031722 4708 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.031730 4708 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.039627 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.132085 4708 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.325896 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.325952 4708 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb" exitCode=137 Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.326006 4708 scope.go:117] "RemoveContainer" containerID="b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.326148 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.339537 4708 scope.go:117] "RemoveContainer" containerID="b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb" Nov 25 05:45:24 crc kubenswrapper[4708]: E1125 05:45:24.339860 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb\": container with ID starting with b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb not found: ID does not exist" containerID="b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.339901 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb"} err="failed to get container status \"b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb\": rpc error: code = NotFound desc = could not find container \"b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb\": container with ID starting with b9029305d818aa78433fb23972b9e96c914fe13966dafbdcfdabd03b732a6ddb not found: ID does not exist" Nov 25 05:45:24 crc kubenswrapper[4708]: I1125 05:45:24.902987 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.287844 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.288879 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerName="controller-manager" containerID="cri-o://c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3" gracePeriod=30 Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.302816 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.303022 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" podUID="4af5429b-294f-4028-8b55-1b55f11d7987" containerName="route-controller-manager" containerID="cri-o://0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73" gracePeriod=30 Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.629829 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.633935 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742255 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqwrj\" (UniqueName: \"kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj\") pod \"bcf83749-f980-4108-b8f5-a618548ccdbf\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742335 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config\") pod \"bcf83749-f980-4108-b8f5-a618548ccdbf\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742372 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca\") pod \"4af5429b-294f-4028-8b55-1b55f11d7987\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742402 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles\") pod \"bcf83749-f980-4108-b8f5-a618548ccdbf\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742445 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert\") pod \"4af5429b-294f-4028-8b55-1b55f11d7987\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742504 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert\") pod \"bcf83749-f980-4108-b8f5-a618548ccdbf\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742570 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config\") pod \"4af5429b-294f-4028-8b55-1b55f11d7987\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742600 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5sg5\" (UniqueName: \"kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5\") pod \"4af5429b-294f-4028-8b55-1b55f11d7987\" (UID: \"4af5429b-294f-4028-8b55-1b55f11d7987\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.742658 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca\") pod \"bcf83749-f980-4108-b8f5-a618548ccdbf\" (UID: \"bcf83749-f980-4108-b8f5-a618548ccdbf\") " Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.743836 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca" (OuterVolumeSpecName: "client-ca") pod "4af5429b-294f-4028-8b55-1b55f11d7987" (UID: "4af5429b-294f-4028-8b55-1b55f11d7987"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.743877 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config" (OuterVolumeSpecName: "config") pod "bcf83749-f980-4108-b8f5-a618548ccdbf" (UID: "bcf83749-f980-4108-b8f5-a618548ccdbf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.743844 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bcf83749-f980-4108-b8f5-a618548ccdbf" (UID: "bcf83749-f980-4108-b8f5-a618548ccdbf"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.743848 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config" (OuterVolumeSpecName: "config") pod "4af5429b-294f-4028-8b55-1b55f11d7987" (UID: "4af5429b-294f-4028-8b55-1b55f11d7987"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.744230 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca" (OuterVolumeSpecName: "client-ca") pod "bcf83749-f980-4108-b8f5-a618548ccdbf" (UID: "bcf83749-f980-4108-b8f5-a618548ccdbf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.749256 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4af5429b-294f-4028-8b55-1b55f11d7987" (UID: "4af5429b-294f-4028-8b55-1b55f11d7987"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.750950 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5" (OuterVolumeSpecName: "kube-api-access-k5sg5") pod "4af5429b-294f-4028-8b55-1b55f11d7987" (UID: "4af5429b-294f-4028-8b55-1b55f11d7987"). InnerVolumeSpecName "kube-api-access-k5sg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.751074 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bcf83749-f980-4108-b8f5-a618548ccdbf" (UID: "bcf83749-f980-4108-b8f5-a618548ccdbf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.751328 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj" (OuterVolumeSpecName: "kube-api-access-fqwrj") pod "bcf83749-f980-4108-b8f5-a618548ccdbf" (UID: "bcf83749-f980-4108-b8f5-a618548ccdbf"). InnerVolumeSpecName "kube-api-access-fqwrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.839458 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f8f5f9555-grfc8"] Nov 25 05:45:41 crc kubenswrapper[4708]: E1125 05:45:41.839839 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38ef401b-8325-4595-9d9e-dd6f0aca7078" containerName="collect-profiles" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.839868 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="38ef401b-8325-4595-9d9e-dd6f0aca7078" containerName="collect-profiles" Nov 25 05:45:41 crc kubenswrapper[4708]: E1125 05:45:41.839889 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerName="controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.839898 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerName="controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: E1125 05:45:41.839914 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4af5429b-294f-4028-8b55-1b55f11d7987" containerName="route-controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.839921 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4af5429b-294f-4028-8b55-1b55f11d7987" containerName="route-controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: E1125 05:45:41.839928 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.839934 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.840071 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerName="controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.840089 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.840099 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="38ef401b-8325-4595-9d9e-dd6f0aca7078" containerName="collect-profiles" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.840125 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4af5429b-294f-4028-8b55-1b55f11d7987" containerName="route-controller-manager" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.840865 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844510 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844563 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844577 4708 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844592 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4af5429b-294f-4028-8b55-1b55f11d7987-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844601 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bcf83749-f980-4108-b8f5-a618548ccdbf-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844611 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4af5429b-294f-4028-8b55-1b55f11d7987-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844619 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5sg5\" (UniqueName: \"kubernetes.io/projected/4af5429b-294f-4028-8b55-1b55f11d7987-kube-api-access-k5sg5\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844629 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bcf83749-f980-4108-b8f5-a618548ccdbf-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.844639 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqwrj\" (UniqueName: \"kubernetes.io/projected/bcf83749-f980-4108-b8f5-a618548ccdbf-kube-api-access-fqwrj\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.846208 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8f5f9555-grfc8"] Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.921992 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.923461 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.942687 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.945038 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-proxy-ca-bundles\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.945095 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-config\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.945133 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef81ea3-ed4a-434c-8443-5314da458f25-serving-cert\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.945164 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-client-ca\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:41 crc kubenswrapper[4708]: I1125 05:45:41.945184 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwwl\" (UniqueName: \"kubernetes.io/projected/bef81ea3-ed4a-434c-8443-5314da458f25-kube-api-access-djwwl\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046064 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef81ea3-ed4a-434c-8443-5314da458f25-serving-cert\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046150 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bb8k\" (UniqueName: \"kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046191 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-client-ca\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046215 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046246 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwwl\" (UniqueName: \"kubernetes.io/projected/bef81ea3-ed4a-434c-8443-5314da458f25-kube-api-access-djwwl\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046396 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-proxy-ca-bundles\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046497 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-config\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046565 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.046595 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.047416 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-client-ca\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.047638 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-proxy-ca-bundles\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.048128 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bef81ea3-ed4a-434c-8443-5314da458f25-config\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.049459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bef81ea3-ed4a-434c-8443-5314da458f25-serving-cert\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.073398 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwwl\" (UniqueName: \"kubernetes.io/projected/bef81ea3-ed4a-434c-8443-5314da458f25-kube-api-access-djwwl\") pod \"controller-manager-f8f5f9555-grfc8\" (UID: \"bef81ea3-ed4a-434c-8443-5314da458f25\") " pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.148378 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bb8k\" (UniqueName: \"kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.148430 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.148530 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.148553 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.149794 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.149868 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.151474 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.154298 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.166622 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bb8k\" (UniqueName: \"kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k\") pod \"route-controller-manager-bf9b5d58f-2gv8p\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.250967 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.309840 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f8f5f9555-grfc8"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.415106 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.417397 4708 generic.go:334] "Generic (PLEG): container finished" podID="4af5429b-294f-4028-8b55-1b55f11d7987" containerID="0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73" exitCode=0 Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.417458 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" event={"ID":"4af5429b-294f-4028-8b55-1b55f11d7987","Type":"ContainerDied","Data":"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73"} Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.417486 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" event={"ID":"4af5429b-294f-4028-8b55-1b55f11d7987","Type":"ContainerDied","Data":"84bc8b687987643fd6b723d22d0db3823288b8049c404bed01d348ae966f3d05"} Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.417505 4708 scope.go:117] "RemoveContainer" containerID="0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.417648 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.423220 4708 generic.go:334] "Generic (PLEG): container finished" podID="bcf83749-f980-4108-b8f5-a618548ccdbf" containerID="c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3" exitCode=0 Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.423270 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" event={"ID":"bcf83749-f980-4108-b8f5-a618548ccdbf","Type":"ContainerDied","Data":"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3"} Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.423291 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" event={"ID":"bcf83749-f980-4108-b8f5-a618548ccdbf","Type":"ContainerDied","Data":"bb629ded7086b72a769a58c1ad26b5427b6ee9dd50ef9e653e1d0521b88b4bf2"} Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.423339 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-6gs2b" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.429598 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" event={"ID":"bef81ea3-ed4a-434c-8443-5314da458f25","Type":"ContainerStarted","Data":"e5b02e8be572333df2cea1b7ca17b141deb31d6e05ed80e604cb605a30c8cced"} Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.436039 4708 scope.go:117] "RemoveContainer" containerID="0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73" Nov 25 05:45:42 crc kubenswrapper[4708]: E1125 05:45:42.436729 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73\": container with ID starting with 0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73 not found: ID does not exist" containerID="0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.436770 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73"} err="failed to get container status \"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73\": rpc error: code = NotFound desc = could not find container \"0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73\": container with ID starting with 0b80ca18d12e01630469507916295d1fc381cef64773fd67c45a67a079394a73 not found: ID does not exist" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.436796 4708 scope.go:117] "RemoveContainer" containerID="c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3" Nov 25 05:45:42 crc kubenswrapper[4708]: W1125 05:45:42.441485 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd66ca80d_57af_4256_95b9_08449a7fa48f.slice/crio-7d9d5343ad67d261ee319c5348210e46993016e60298f263274b3d17eb31d1f7 WatchSource:0}: Error finding container 7d9d5343ad67d261ee319c5348210e46993016e60298f263274b3d17eb31d1f7: Status 404 returned error can't find the container with id 7d9d5343ad67d261ee319c5348210e46993016e60298f263274b3d17eb31d1f7 Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.501591 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.502592 4708 scope.go:117] "RemoveContainer" containerID="c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3" Nov 25 05:45:42 crc kubenswrapper[4708]: E1125 05:45:42.503056 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3\": container with ID starting with c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3 not found: ID does not exist" containerID="c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.503088 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3"} err="failed to get container status \"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3\": rpc error: code = NotFound desc = could not find container \"c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3\": container with ID starting with c51445f67b4599438d8164237d31da8970aea1ce44db8a592498108c4da67fe3 not found: ID does not exist" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.504634 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-gp6rv"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.514392 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.520450 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-6gs2b"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.841050 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.898328 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4af5429b-294f-4028-8b55-1b55f11d7987" path="/var/lib/kubelet/pods/4af5429b-294f-4028-8b55-1b55f11d7987/volumes" Nov 25 05:45:42 crc kubenswrapper[4708]: I1125 05:45:42.899055 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcf83749-f980-4108-b8f5-a618548ccdbf" path="/var/lib/kubelet/pods/bcf83749-f980-4108-b8f5-a618548ccdbf/volumes" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.439986 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" event={"ID":"bef81ea3-ed4a-434c-8443-5314da458f25","Type":"ContainerStarted","Data":"1f27142e76686ff9b26bcd64a349f402c485ee443c8a6650ad65402b4c38ac03"} Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.440104 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.444788 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" event={"ID":"d66ca80d-57af-4256-95b9-08449a7fa48f","Type":"ContainerStarted","Data":"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354"} Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.444834 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" event={"ID":"d66ca80d-57af-4256-95b9-08449a7fa48f","Type":"ContainerStarted","Data":"7d9d5343ad67d261ee319c5348210e46993016e60298f263274b3d17eb31d1f7"} Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.445097 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.445173 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.449221 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.460730 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f8f5f9555-grfc8" podStartSLOduration=2.460707029 podStartE2EDuration="2.460707029s" podCreationTimestamp="2025-11-25 05:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:45:43.458804472 +0000 UTC m=+284.867637859" watchObservedRunningTime="2025-11-25 05:45:43.460707029 +0000 UTC m=+284.869540415" Nov 25 05:45:43 crc kubenswrapper[4708]: I1125 05:45:43.471694 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" podStartSLOduration=2.471683409 podStartE2EDuration="2.471683409s" podCreationTimestamp="2025-11-25 05:45:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:45:43.47054839 +0000 UTC m=+284.879381776" watchObservedRunningTime="2025-11-25 05:45:43.471683409 +0000 UTC m=+284.880516795" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.451179 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" podUID="d66ca80d-57af-4256-95b9-08449a7fa48f" containerName="route-controller-manager" containerID="cri-o://8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354" gracePeriod=30 Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.765418 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.784343 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert\") pod \"d66ca80d-57af-4256-95b9-08449a7fa48f\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.784483 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bb8k\" (UniqueName: \"kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k\") pod \"d66ca80d-57af-4256-95b9-08449a7fa48f\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.784568 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config\") pod \"d66ca80d-57af-4256-95b9-08449a7fa48f\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.785604 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config" (OuterVolumeSpecName: "config") pod "d66ca80d-57af-4256-95b9-08449a7fa48f" (UID: "d66ca80d-57af-4256-95b9-08449a7fa48f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.786106 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:45:44 crc kubenswrapper[4708]: E1125 05:45:44.795202 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d66ca80d-57af-4256-95b9-08449a7fa48f" containerName="route-controller-manager" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.795284 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d66ca80d-57af-4256-95b9-08449a7fa48f" containerName="route-controller-manager" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.795996 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d66ca80d-57af-4256-95b9-08449a7fa48f" containerName="route-controller-manager" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.797167 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.799270 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k" (OuterVolumeSpecName: "kube-api-access-9bb8k") pod "d66ca80d-57af-4256-95b9-08449a7fa48f" (UID: "d66ca80d-57af-4256-95b9-08449a7fa48f"). InnerVolumeSpecName "kube-api-access-9bb8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.813010 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d66ca80d-57af-4256-95b9-08449a7fa48f" (UID: "d66ca80d-57af-4256-95b9-08449a7fa48f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.819161 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.885866 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca\") pod \"d66ca80d-57af-4256-95b9-08449a7fa48f\" (UID: \"d66ca80d-57af-4256-95b9-08449a7fa48f\") " Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886320 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca" (OuterVolumeSpecName: "client-ca") pod "d66ca80d-57af-4256-95b9-08449a7fa48f" (UID: "d66ca80d-57af-4256-95b9-08449a7fa48f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886535 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886597 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886696 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x92mb\" (UniqueName: \"kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886784 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886951 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886977 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d66ca80d-57af-4256-95b9-08449a7fa48f-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.886995 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d66ca80d-57af-4256-95b9-08449a7fa48f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.887007 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bb8k\" (UniqueName: \"kubernetes.io/projected/d66ca80d-57af-4256-95b9-08449a7fa48f-kube-api-access-9bb8k\") on node \"crc\" DevicePath \"\"" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.988051 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.988227 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x92mb\" (UniqueName: \"kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.988271 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.988360 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.989281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.989501 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:44 crc kubenswrapper[4708]: I1125 05:45:44.991855 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.002602 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x92mb\" (UniqueName: \"kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb\") pod \"route-controller-manager-7447957dcb-mzxfn\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.122600 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.456095 4708 generic.go:334] "Generic (PLEG): container finished" podID="d66ca80d-57af-4256-95b9-08449a7fa48f" containerID="8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354" exitCode=0 Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.456166 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.456189 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" event={"ID":"d66ca80d-57af-4256-95b9-08449a7fa48f","Type":"ContainerDied","Data":"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354"} Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.456483 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p" event={"ID":"d66ca80d-57af-4256-95b9-08449a7fa48f","Type":"ContainerDied","Data":"7d9d5343ad67d261ee319c5348210e46993016e60298f263274b3d17eb31d1f7"} Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.456502 4708 scope.go:117] "RemoveContainer" containerID="8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.472996 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.474031 4708 scope.go:117] "RemoveContainer" containerID="8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354" Nov 25 05:45:45 crc kubenswrapper[4708]: E1125 05:45:45.474598 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354\": container with ID starting with 8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354 not found: ID does not exist" containerID="8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.474655 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354"} err="failed to get container status \"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354\": rpc error: code = NotFound desc = could not find container \"8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354\": container with ID starting with 8fcba84b7182d80f73f1ae968f174946b8b33111ce73f119e0894205d123b354 not found: ID does not exist" Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.476134 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:45:45 crc kubenswrapper[4708]: I1125 05:45:45.478809 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-2gv8p"] Nov 25 05:45:45 crc kubenswrapper[4708]: W1125 05:45:45.479884 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2691ac53_50f3_46b7_866e_32f9154ce3de.slice/crio-4c0ca711cd4e5ae47b523788b4f2b786032ab1ea55d5bf6549e35a1118afad49 WatchSource:0}: Error finding container 4c0ca711cd4e5ae47b523788b4f2b786032ab1ea55d5bf6549e35a1118afad49: Status 404 returned error can't find the container with id 4c0ca711cd4e5ae47b523788b4f2b786032ab1ea55d5bf6549e35a1118afad49 Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.463457 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" event={"ID":"2691ac53-50f3-46b7-866e-32f9154ce3de","Type":"ContainerStarted","Data":"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b"} Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.463746 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.463759 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" event={"ID":"2691ac53-50f3-46b7-866e-32f9154ce3de","Type":"ContainerStarted","Data":"4c0ca711cd4e5ae47b523788b4f2b786032ab1ea55d5bf6549e35a1118afad49"} Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.468622 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.478893 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" podStartSLOduration=4.478879999 podStartE2EDuration="4.478879999s" podCreationTimestamp="2025-11-25 05:45:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:45:46.475218626 +0000 UTC m=+287.884052012" watchObservedRunningTime="2025-11-25 05:45:46.478879999 +0000 UTC m=+287.887713385" Nov 25 05:45:46 crc kubenswrapper[4708]: I1125 05:45:46.899692 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d66ca80d-57af-4256-95b9-08449a7fa48f" path="/var/lib/kubelet/pods/d66ca80d-57af-4256-95b9-08449a7fa48f/volumes" Nov 25 05:46:33 crc kubenswrapper[4708]: I1125 05:46:33.951439 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bwxvs"] Nov 25 05:46:33 crc kubenswrapper[4708]: I1125 05:46:33.952866 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:33 crc kubenswrapper[4708]: I1125 05:46:33.958001 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bwxvs"] Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112017 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00f47aee-85b8-4223-bc22-d5ca111e4cf2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112088 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-trusted-ca\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112251 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-bound-sa-token\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112369 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdjrp\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-kube-api-access-jdjrp\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112459 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112490 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-certificates\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112740 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-tls\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.112792 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00f47aee-85b8-4223-bc22-d5ca111e4cf2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.131226 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214112 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-tls\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214154 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00f47aee-85b8-4223-bc22-d5ca111e4cf2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214198 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00f47aee-85b8-4223-bc22-d5ca111e4cf2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214222 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-trusted-ca\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214247 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-bound-sa-token\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214271 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdjrp\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-kube-api-access-jdjrp\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214302 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-certificates\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.214781 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/00f47aee-85b8-4223-bc22-d5ca111e4cf2-ca-trust-extracted\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.215459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-certificates\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.215750 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00f47aee-85b8-4223-bc22-d5ca111e4cf2-trusted-ca\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.219989 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-registry-tls\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.220161 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/00f47aee-85b8-4223-bc22-d5ca111e4cf2-installation-pull-secrets\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.228566 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-bound-sa-token\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.230032 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdjrp\" (UniqueName: \"kubernetes.io/projected/00f47aee-85b8-4223-bc22-d5ca111e4cf2-kube-api-access-jdjrp\") pod \"image-registry-66df7c8f76-bwxvs\" (UID: \"00f47aee-85b8-4223-bc22-d5ca111e4cf2\") " pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.269981 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.629606 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-bwxvs"] Nov 25 05:46:34 crc kubenswrapper[4708]: I1125 05:46:34.677782 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" event={"ID":"00f47aee-85b8-4223-bc22-d5ca111e4cf2","Type":"ContainerStarted","Data":"686f142adbaf7e136b81bfce96e182db6c2d8dabdfece00bab0bce5127563eb5"} Nov 25 05:46:35 crc kubenswrapper[4708]: I1125 05:46:35.685894 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" event={"ID":"00f47aee-85b8-4223-bc22-d5ca111e4cf2","Type":"ContainerStarted","Data":"8ee0548ecfa82b78db01dda15b2230a73771cd521f254f8ad081e7c9530e4057"} Nov 25 05:46:35 crc kubenswrapper[4708]: I1125 05:46:35.686230 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:35 crc kubenswrapper[4708]: I1125 05:46:35.703535 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" podStartSLOduration=2.703505354 podStartE2EDuration="2.703505354s" podCreationTimestamp="2025-11-25 05:46:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:46:35.701033523 +0000 UTC m=+337.109866909" watchObservedRunningTime="2025-11-25 05:46:35.703505354 +0000 UTC m=+337.112338740" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.240670 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.241315 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" podUID="2691ac53-50f3-46b7-866e-32f9154ce3de" containerName="route-controller-manager" containerID="cri-o://ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b" gracePeriod=30 Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.564655 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.710672 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca\") pod \"2691ac53-50f3-46b7-866e-32f9154ce3de\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.710724 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert\") pod \"2691ac53-50f3-46b7-866e-32f9154ce3de\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.710772 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x92mb\" (UniqueName: \"kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb\") pod \"2691ac53-50f3-46b7-866e-32f9154ce3de\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.711645 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config\") pod \"2691ac53-50f3-46b7-866e-32f9154ce3de\" (UID: \"2691ac53-50f3-46b7-866e-32f9154ce3de\") " Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.711964 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca" (OuterVolumeSpecName: "client-ca") pod "2691ac53-50f3-46b7-866e-32f9154ce3de" (UID: "2691ac53-50f3-46b7-866e-32f9154ce3de"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.712537 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config" (OuterVolumeSpecName: "config") pod "2691ac53-50f3-46b7-866e-32f9154ce3de" (UID: "2691ac53-50f3-46b7-866e-32f9154ce3de"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.714270 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.714438 4708 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2691ac53-50f3-46b7-866e-32f9154ce3de-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.718472 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb" (OuterVolumeSpecName: "kube-api-access-x92mb") pod "2691ac53-50f3-46b7-866e-32f9154ce3de" (UID: "2691ac53-50f3-46b7-866e-32f9154ce3de"). InnerVolumeSpecName "kube-api-access-x92mb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.719806 4708 generic.go:334] "Generic (PLEG): container finished" podID="2691ac53-50f3-46b7-866e-32f9154ce3de" containerID="ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b" exitCode=0 Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.719854 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" event={"ID":"2691ac53-50f3-46b7-866e-32f9154ce3de","Type":"ContainerDied","Data":"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b"} Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.719913 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.719933 4708 scope.go:117] "RemoveContainer" containerID="ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.719920 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn" event={"ID":"2691ac53-50f3-46b7-866e-32f9154ce3de","Type":"ContainerDied","Data":"4c0ca711cd4e5ae47b523788b4f2b786032ab1ea55d5bf6549e35a1118afad49"} Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.723407 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2691ac53-50f3-46b7-866e-32f9154ce3de" (UID: "2691ac53-50f3-46b7-866e-32f9154ce3de"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.740047 4708 scope.go:117] "RemoveContainer" containerID="ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b" Nov 25 05:46:41 crc kubenswrapper[4708]: E1125 05:46:41.740351 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b\": container with ID starting with ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b not found: ID does not exist" containerID="ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.740453 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b"} err="failed to get container status \"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b\": rpc error: code = NotFound desc = could not find container \"ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b\": container with ID starting with ee92701f71e826066fb975e5571244d59faeaf4f88765ba75ca5d1d88c08788b not found: ID does not exist" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.815078 4708 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2691ac53-50f3-46b7-866e-32f9154ce3de-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:46:41 crc kubenswrapper[4708]: I1125 05:46:41.815167 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x92mb\" (UniqueName: \"kubernetes.io/projected/2691ac53-50f3-46b7-866e-32f9154ce3de-kube-api-access-x92mb\") on node \"crc\" DevicePath \"\"" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.044558 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.048950 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7447957dcb-mzxfn"] Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.615645 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv"] Nov 25 05:46:42 crc kubenswrapper[4708]: E1125 05:46:42.615859 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2691ac53-50f3-46b7-866e-32f9154ce3de" containerName="route-controller-manager" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.615871 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2691ac53-50f3-46b7-866e-32f9154ce3de" containerName="route-controller-manager" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.615979 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="2691ac53-50f3-46b7-866e-32f9154ce3de" containerName="route-controller-manager" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.616394 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.617848 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.617881 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.618453 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.618630 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.618628 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.618734 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.622724 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-client-ca\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.622854 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-config\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.622935 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-serving-cert\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.623000 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rzq6\" (UniqueName: \"kubernetes.io/projected/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-kube-api-access-2rzq6\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.624513 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv"] Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.723970 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-client-ca\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.724044 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-config\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.724102 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-serving-cert\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.724149 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rzq6\" (UniqueName: \"kubernetes.io/projected/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-kube-api-access-2rzq6\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.725128 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-client-ca\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.726014 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-config\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.729433 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-serving-cert\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.737544 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rzq6\" (UniqueName: \"kubernetes.io/projected/f06ebe69-0237-4c1c-8d25-b9d0306d4c30-kube-api-access-2rzq6\") pod \"route-controller-manager-bf9b5d58f-9d7pv\" (UID: \"f06ebe69-0237-4c1c-8d25-b9d0306d4c30\") " pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.898030 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2691ac53-50f3-46b7-866e-32f9154ce3de" path="/var/lib/kubelet/pods/2691ac53-50f3-46b7-866e-32f9154ce3de/volumes" Nov 25 05:46:42 crc kubenswrapper[4708]: I1125 05:46:42.930118 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.279781 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv"] Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.730067 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" event={"ID":"f06ebe69-0237-4c1c-8d25-b9d0306d4c30","Type":"ContainerStarted","Data":"8fb03e27ffbe1a1ff934b7290673e5221d2077b1cf62a218f5438f152d559a39"} Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.730408 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" event={"ID":"f06ebe69-0237-4c1c-8d25-b9d0306d4c30","Type":"ContainerStarted","Data":"ea84ec6d80bdb693453ed245611cf17d40f0628dba185344343f4ea79381dd07"} Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.730425 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.743155 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" podStartSLOduration=2.743143703 podStartE2EDuration="2.743143703s" podCreationTimestamp="2025-11-25 05:46:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:46:43.740895474 +0000 UTC m=+345.149728861" watchObservedRunningTime="2025-11-25 05:46:43.743143703 +0000 UTC m=+345.151977090" Nov 25 05:46:43 crc kubenswrapper[4708]: I1125 05:46:43.814445 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bf9b5d58f-9d7pv" Nov 25 05:46:44 crc kubenswrapper[4708]: I1125 05:46:44.357476 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:46:44 crc kubenswrapper[4708]: I1125 05:46:44.357556 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:46:54 crc kubenswrapper[4708]: I1125 05:46:54.276378 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-bwxvs" Nov 25 05:46:54 crc kubenswrapper[4708]: I1125 05:46:54.318117 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:47:14 crc kubenswrapper[4708]: I1125 05:47:14.357300 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:47:14 crc kubenswrapper[4708]: I1125 05:47:14.357871 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.345887 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" podUID="e27e8c69-3b12-4193-9fc3-26d339707d95" containerName="registry" containerID="cri-o://f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16" gracePeriod=30 Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.644640 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.833951 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvwgt\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834002 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834035 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834073 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834231 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834266 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834314 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.834350 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates\") pod \"e27e8c69-3b12-4193-9fc3-26d339707d95\" (UID: \"e27e8c69-3b12-4193-9fc3-26d339707d95\") " Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.835313 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.835685 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.840208 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt" (OuterVolumeSpecName: "kube-api-access-zvwgt") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "kube-api-access-zvwgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.840570 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.840754 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.842089 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.842764 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.852842 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "e27e8c69-3b12-4193-9fc3-26d339707d95" (UID: "e27e8c69-3b12-4193-9fc3-26d339707d95"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.896823 4708 generic.go:334] "Generic (PLEG): container finished" podID="e27e8c69-3b12-4193-9fc3-26d339707d95" containerID="f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16" exitCode=0 Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.896870 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" event={"ID":"e27e8c69-3b12-4193-9fc3-26d339707d95","Type":"ContainerDied","Data":"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16"} Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.896920 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" event={"ID":"e27e8c69-3b12-4193-9fc3-26d339707d95","Type":"ContainerDied","Data":"a76c9f9d3a4490900ba7f87279bdb08b99883fa71294df5f635bc5fe77cfd076"} Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.896916 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hmtsm" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.897007 4708 scope.go:117] "RemoveContainer" containerID="f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.910069 4708 scope.go:117] "RemoveContainer" containerID="f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16" Nov 25 05:47:19 crc kubenswrapper[4708]: E1125 05:47:19.910381 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16\": container with ID starting with f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16 not found: ID does not exist" containerID="f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.910414 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16"} err="failed to get container status \"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16\": rpc error: code = NotFound desc = could not find container \"f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16\": container with ID starting with f0e5c78cb4f30689d7750ce649993c5945cffb22291ebbd837de0baa1b1fec16 not found: ID does not exist" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.917132 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.920227 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hmtsm"] Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935610 4708 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935635 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvwgt\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-kube-api-access-zvwgt\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935646 4708 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935656 4708 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/e27e8c69-3b12-4193-9fc3-26d339707d95-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935666 4708 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/e27e8c69-3b12-4193-9fc3-26d339707d95-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935674 4708 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e27e8c69-3b12-4193-9fc3-26d339707d95-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:19 crc kubenswrapper[4708]: I1125 05:47:19.935682 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e27e8c69-3b12-4193-9fc3-26d339707d95-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:47:20 crc kubenswrapper[4708]: I1125 05:47:20.900781 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e27e8c69-3b12-4193-9fc3-26d339707d95" path="/var/lib/kubelet/pods/e27e8c69-3b12-4193-9fc3-26d339707d95/volumes" Nov 25 05:47:44 crc kubenswrapper[4708]: I1125 05:47:44.357385 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:47:44 crc kubenswrapper[4708]: I1125 05:47:44.357949 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:47:44 crc kubenswrapper[4708]: I1125 05:47:44.357992 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:47:44 crc kubenswrapper[4708]: I1125 05:47:44.358647 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:47:44 crc kubenswrapper[4708]: I1125 05:47:44.358703 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2" gracePeriod=600 Nov 25 05:47:45 crc kubenswrapper[4708]: I1125 05:47:45.000722 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2" exitCode=0 Nov 25 05:47:45 crc kubenswrapper[4708]: I1125 05:47:45.001040 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2"} Nov 25 05:47:45 crc kubenswrapper[4708]: I1125 05:47:45.001070 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f"} Nov 25 05:47:45 crc kubenswrapper[4708]: I1125 05:47:45.001089 4708 scope.go:117] "RemoveContainer" containerID="bbc54caf22d2d89986cb76b20b27f1c97b98dc9b2470e5149c8f0e0c45307545" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.957724 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t4xc2"] Nov 25 05:48:46 crc kubenswrapper[4708]: E1125 05:48:46.958697 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e27e8c69-3b12-4193-9fc3-26d339707d95" containerName="registry" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.958717 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e27e8c69-3b12-4193-9fc3-26d339707d95" containerName="registry" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.958827 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e27e8c69-3b12-4193-9fc3-26d339707d95" containerName="registry" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.959357 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.974639 4708 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5zh42" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.975020 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.975163 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.980814 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pfp8b"] Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.981973 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pfp8b" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.983510 4708 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-l7lh2" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.986149 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t4xc2"] Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.992768 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7dc5l"] Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.994752 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.996682 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7dc5l"] Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.997197 4708 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-l8lrp" Nov 25 05:48:46 crc kubenswrapper[4708]: I1125 05:48:46.999635 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pfp8b"] Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.071088 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2q74\" (UniqueName: \"kubernetes.io/projected/6cd001d7-f756-4a83-8d4b-0bc9c5198b0a-kube-api-access-l2q74\") pod \"cert-manager-cainjector-7f985d654d-t4xc2\" (UID: \"6cd001d7-f756-4a83-8d4b-0bc9c5198b0a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.172560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcgwj\" (UniqueName: \"kubernetes.io/projected/8775bc9d-3d1d-471a-b6af-c057f25a7f12-kube-api-access-vcgwj\") pod \"cert-manager-5b446d88c5-pfp8b\" (UID: \"8775bc9d-3d1d-471a-b6af-c057f25a7f12\") " pod="cert-manager/cert-manager-5b446d88c5-pfp8b" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.172972 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2q74\" (UniqueName: \"kubernetes.io/projected/6cd001d7-f756-4a83-8d4b-0bc9c5198b0a-kube-api-access-l2q74\") pod \"cert-manager-cainjector-7f985d654d-t4xc2\" (UID: \"6cd001d7-f756-4a83-8d4b-0bc9c5198b0a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.173061 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgq8c\" (UniqueName: \"kubernetes.io/projected/d043f20a-ee01-40c7-86a5-ff37f52d7f07-kube-api-access-lgq8c\") pod \"cert-manager-webhook-5655c58dd6-7dc5l\" (UID: \"d043f20a-ee01-40c7-86a5-ff37f52d7f07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.190748 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2q74\" (UniqueName: \"kubernetes.io/projected/6cd001d7-f756-4a83-8d4b-0bc9c5198b0a-kube-api-access-l2q74\") pod \"cert-manager-cainjector-7f985d654d-t4xc2\" (UID: \"6cd001d7-f756-4a83-8d4b-0bc9c5198b0a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.274162 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgq8c\" (UniqueName: \"kubernetes.io/projected/d043f20a-ee01-40c7-86a5-ff37f52d7f07-kube-api-access-lgq8c\") pod \"cert-manager-webhook-5655c58dd6-7dc5l\" (UID: \"d043f20a-ee01-40c7-86a5-ff37f52d7f07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.274266 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcgwj\" (UniqueName: \"kubernetes.io/projected/8775bc9d-3d1d-471a-b6af-c057f25a7f12-kube-api-access-vcgwj\") pod \"cert-manager-5b446d88c5-pfp8b\" (UID: \"8775bc9d-3d1d-471a-b6af-c057f25a7f12\") " pod="cert-manager/cert-manager-5b446d88c5-pfp8b" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.274287 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.289401 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcgwj\" (UniqueName: \"kubernetes.io/projected/8775bc9d-3d1d-471a-b6af-c057f25a7f12-kube-api-access-vcgwj\") pod \"cert-manager-5b446d88c5-pfp8b\" (UID: \"8775bc9d-3d1d-471a-b6af-c057f25a7f12\") " pod="cert-manager/cert-manager-5b446d88c5-pfp8b" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.291791 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgq8c\" (UniqueName: \"kubernetes.io/projected/d043f20a-ee01-40c7-86a5-ff37f52d7f07-kube-api-access-lgq8c\") pod \"cert-manager-webhook-5655c58dd6-7dc5l\" (UID: \"d043f20a-ee01-40c7-86a5-ff37f52d7f07\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.295693 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-pfp8b" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.304352 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.647949 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-t4xc2"] Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.657512 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.683328 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-pfp8b"] Nov 25 05:48:47 crc kubenswrapper[4708]: W1125 05:48:47.687532 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8775bc9d_3d1d_471a_b6af_c057f25a7f12.slice/crio-43b78e00fae74f7d70a3fab934e9d7e785cfc89c4cfce51f3d3e91b421b2326f WatchSource:0}: Error finding container 43b78e00fae74f7d70a3fab934e9d7e785cfc89c4cfce51f3d3e91b421b2326f: Status 404 returned error can't find the container with id 43b78e00fae74f7d70a3fab934e9d7e785cfc89c4cfce51f3d3e91b421b2326f Nov 25 05:48:47 crc kubenswrapper[4708]: I1125 05:48:47.705763 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-7dc5l"] Nov 25 05:48:47 crc kubenswrapper[4708]: W1125 05:48:47.709155 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd043f20a_ee01_40c7_86a5_ff37f52d7f07.slice/crio-0e213fbfab9a5b40e05ba73289ba09b324f1e1e2913d3d19c569490d76a9f2c9 WatchSource:0}: Error finding container 0e213fbfab9a5b40e05ba73289ba09b324f1e1e2913d3d19c569490d76a9f2c9: Status 404 returned error can't find the container with id 0e213fbfab9a5b40e05ba73289ba09b324f1e1e2913d3d19c569490d76a9f2c9 Nov 25 05:48:48 crc kubenswrapper[4708]: I1125 05:48:48.250711 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" event={"ID":"d043f20a-ee01-40c7-86a5-ff37f52d7f07","Type":"ContainerStarted","Data":"0e213fbfab9a5b40e05ba73289ba09b324f1e1e2913d3d19c569490d76a9f2c9"} Nov 25 05:48:48 crc kubenswrapper[4708]: I1125 05:48:48.252551 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" event={"ID":"6cd001d7-f756-4a83-8d4b-0bc9c5198b0a","Type":"ContainerStarted","Data":"d54c565328d3e233608a13fc62f2f6b5fe95be241acce1af3751981b4fe6809d"} Nov 25 05:48:48 crc kubenswrapper[4708]: I1125 05:48:48.253537 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pfp8b" event={"ID":"8775bc9d-3d1d-471a-b6af-c057f25a7f12","Type":"ContainerStarted","Data":"43b78e00fae74f7d70a3fab934e9d7e785cfc89c4cfce51f3d3e91b421b2326f"} Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.271947 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" event={"ID":"d043f20a-ee01-40c7-86a5-ff37f52d7f07","Type":"ContainerStarted","Data":"4734b70fa2b09ab0858d73bbca2756bae0d57b094983e2257d8f18e19f238696"} Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.272632 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.274550 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" event={"ID":"6cd001d7-f756-4a83-8d4b-0bc9c5198b0a","Type":"ContainerStarted","Data":"0064cfdf0b83886f02723ee856bc1fc1666218189da7ec8f0240c989b5a47cc5"} Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.275666 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-pfp8b" event={"ID":"8775bc9d-3d1d-471a-b6af-c057f25a7f12","Type":"ContainerStarted","Data":"bf81829310f796d7c09a2789f7a872165863bc80d3c10dc51ded8c56a38bf50c"} Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.283349 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" podStartSLOduration=2.79081299 podStartE2EDuration="5.283337218s" podCreationTimestamp="2025-11-25 05:48:46 +0000 UTC" firstStartedPulling="2025-11-25 05:48:47.710875175 +0000 UTC m=+469.119708562" lastFinishedPulling="2025-11-25 05:48:50.203399403 +0000 UTC m=+471.612232790" observedRunningTime="2025-11-25 05:48:51.282857473 +0000 UTC m=+472.691690860" watchObservedRunningTime="2025-11-25 05:48:51.283337218 +0000 UTC m=+472.692170604" Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.292833 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-t4xc2" podStartSLOduration=2.747478432 podStartE2EDuration="5.292814933s" podCreationTimestamp="2025-11-25 05:48:46 +0000 UTC" firstStartedPulling="2025-11-25 05:48:47.657215986 +0000 UTC m=+469.066049372" lastFinishedPulling="2025-11-25 05:48:50.202552497 +0000 UTC m=+471.611385873" observedRunningTime="2025-11-25 05:48:51.292157624 +0000 UTC m=+472.700991010" watchObservedRunningTime="2025-11-25 05:48:51.292814933 +0000 UTC m=+472.701648319" Nov 25 05:48:51 crc kubenswrapper[4708]: I1125 05:48:51.312242 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-pfp8b" podStartSLOduration=2.803815851 podStartE2EDuration="5.312217868s" podCreationTimestamp="2025-11-25 05:48:46 +0000 UTC" firstStartedPulling="2025-11-25 05:48:47.689944701 +0000 UTC m=+469.098778087" lastFinishedPulling="2025-11-25 05:48:50.198346729 +0000 UTC m=+471.607180104" observedRunningTime="2025-11-25 05:48:51.303505545 +0000 UTC m=+472.712338921" watchObservedRunningTime="2025-11-25 05:48:51.312217868 +0000 UTC m=+472.721051255" Nov 25 05:48:57 crc kubenswrapper[4708]: I1125 05:48:57.308262 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-7dc5l" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646448 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qppcd"] Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646832 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-controller" containerID="cri-o://62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646893 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="nbdb" containerID="cri-o://98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646925 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="northd" containerID="cri-o://b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646942 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.646999 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="sbdb" containerID="cri-o://b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.647054 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-acl-logging" containerID="cri-o://f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.647109 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-node" containerID="cri-o://59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.670726 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" containerID="cri-o://672e0618a41d664b8375415fd94c0799d59202a4ee97a3f1acfa4a3d42de6ec2" gracePeriod=30 Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.903860 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovnkube-controller/3.log" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.906084 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovn-acl-logging/0.log" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.906538 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qppcd_6d78be28-7e93-4fda-98cd-521398f9b3e4/ovn-controller/0.log" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.906935 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909257 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7wnr\" (UniqueName: \"kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909289 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909327 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909352 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909367 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909388 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909534 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909556 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log" (OuterVolumeSpecName: "node-log") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909590 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909643 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909735 4708 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909754 4708 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909764 4708 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.909773 4708 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.913920 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr" (OuterVolumeSpecName: "kube-api-access-q7wnr") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "kube-api-access-q7wnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.914738 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949322 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rvg5p"] Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949604 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="northd" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949616 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="northd" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949632 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-acl-logging" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949637 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-acl-logging" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949644 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-node" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949651 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-node" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949659 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949664 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949675 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="sbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949681 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="sbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949691 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949699 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949705 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949710 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949716 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949722 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949729 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949736 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949748 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kubecfg-setup" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949754 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kubecfg-setup" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949760 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="nbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949766 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="nbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.949773 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949778 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949884 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949894 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-acl-logging" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949903 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949910 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949918 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="nbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949925 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="sbdb" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949932 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949938 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="northd" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949946 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovn-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949952 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.949959 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="kube-rbac-proxy-node" Nov 25 05:48:58 crc kubenswrapper[4708]: E1125 05:48:58.950054 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.950061 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.950156 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" containerName="ovnkube-controller" Nov 25 05:48:58 crc kubenswrapper[4708]: I1125 05:48:58.951706 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010571 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010603 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010645 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010672 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010685 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010724 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010692 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010758 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010774 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010792 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010809 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010811 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010830 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010851 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010853 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010872 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010877 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010895 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010946 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.010960 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin\") pod \"6d78be28-7e93-4fda-98cd-521398f9b3e4\" (UID: \"6d78be28-7e93-4fda-98cd-521398f9b3e4\") " Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011115 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-kubelet\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011133 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011141 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011151 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89b7t\" (UniqueName: \"kubernetes.io/projected/81ff213f-6eda-4128-bee4-8f428c972f61-kube-api-access-89b7t\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011174 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket" (OuterVolumeSpecName: "log-socket") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011179 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash" (OuterVolumeSpecName: "host-slash") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011194 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011207 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011227 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-log-socket\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011268 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-slash\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011292 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-bin\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011307 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-env-overrides\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011337 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-config\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011360 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-ovn\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011465 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-systemd-units\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011491 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-etc-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011534 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011544 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011590 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-netd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011617 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-netns\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011602 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011640 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-var-lib-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011662 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-systemd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011677 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-script-lib\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011773 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/81ff213f-6eda-4128-bee4-8f428c972f61-ovn-node-metrics-cert\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011815 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011847 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-node-log\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011949 4708 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011970 4708 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011979 4708 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011988 4708 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.011998 4708 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012011 4708 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012021 4708 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012030 4708 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012037 4708 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012046 4708 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012054 4708 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012062 4708 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012071 4708 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012080 4708 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6d78be28-7e93-4fda-98cd-521398f9b3e4-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.012090 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7wnr\" (UniqueName: \"kubernetes.io/projected/6d78be28-7e93-4fda-98cd-521398f9b3e4-kube-api-access-q7wnr\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.023778 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6d78be28-7e93-4fda-98cd-521398f9b3e4" (UID: "6d78be28-7e93-4fda-98cd-521398f9b3e4"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.063665 4708 scope.go:117] "RemoveContainer" containerID="b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.080260 4708 scope.go:117] "RemoveContainer" containerID="62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.096253 4708 scope.go:117] "RemoveContainer" containerID="98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.108825 4708 scope.go:117] "RemoveContainer" containerID="b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113695 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-slash\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113738 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-bin\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113761 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-env-overrides\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113789 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-config\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113821 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-ovn\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113822 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-bin\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113851 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-systemd-units\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113874 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113822 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-slash\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113895 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-etc-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113926 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-netd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113927 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-ovn\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113945 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-netns\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113967 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-var-lib-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113974 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.113989 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-systemd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114014 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-systemd-units\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-script-lib\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114107 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/81ff213f-6eda-4128-bee4-8f428c972f61-ovn-node-metrics-cert\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114145 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114157 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-netns\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114184 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-node-log\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114201 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-etc-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114209 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-kubelet\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114237 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114261 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89b7t\" (UniqueName: \"kubernetes.io/projected/81ff213f-6eda-4128-bee4-8f428c972f61-kube-api-access-89b7t\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114311 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114236 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-kubelet\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114392 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-run-systemd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114431 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-node-log\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114468 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-run-ovn-kubernetes\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114485 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-var-lib-openvswitch\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114544 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-log-socket\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114510 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-log-socket\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114613 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-config\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114753 4708 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6d78be28-7e93-4fda-98cd-521398f9b3e4-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114884 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-env-overrides\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.115095 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/81ff213f-6eda-4128-bee4-8f428c972f61-ovnkube-script-lib\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.114614 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/81ff213f-6eda-4128-bee4-8f428c972f61-host-cni-netd\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.117409 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/81ff213f-6eda-4128-bee4-8f428c972f61-ovn-node-metrics-cert\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.121648 4708 scope.go:117] "RemoveContainer" containerID="b6c6f9946bf42048f3b46724e0262446edc05ad6341470292c1d30cedca88f9b" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.132132 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89b7t\" (UniqueName: \"kubernetes.io/projected/81ff213f-6eda-4128-bee4-8f428c972f61-kube-api-access-89b7t\") pod \"ovnkube-node-rvg5p\" (UID: \"81ff213f-6eda-4128-bee4-8f428c972f61\") " pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.137195 4708 scope.go:117] "RemoveContainer" containerID="1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.150019 4708 scope.go:117] "RemoveContainer" containerID="59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.161396 4708 scope.go:117] "RemoveContainer" containerID="f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.171985 4708 scope.go:117] "RemoveContainer" containerID="ced2bd1360b79b035493531df3f564ca3ed0119fbee09c4b6961184bb9e4e7a6" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.184182 4708 scope.go:117] "RemoveContainer" containerID="672e0618a41d664b8375415fd94c0799d59202a4ee97a3f1acfa4a3d42de6ec2" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.194221 4708 scope.go:117] "RemoveContainer" containerID="cee8602b7b471af8c077b0a9afdc912db9bfd5933a69fee30cac9941b22093b1" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.264342 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:48:59 crc kubenswrapper[4708]: W1125 05:48:59.279208 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81ff213f_6eda_4128_bee4_8f428c972f61.slice/crio-6cfd7878429b77d861de55f6401fdda7fd677eff48ebc6dfa94ec6aa883f4855 WatchSource:0}: Error finding container 6cfd7878429b77d861de55f6401fdda7fd677eff48ebc6dfa94ec6aa883f4855: Status 404 returned error can't find the container with id 6cfd7878429b77d861de55f6401fdda7fd677eff48ebc6dfa94ec6aa883f4855 Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.313471 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rjrnw_abcfd1aa-ec01-4cae-8858-0d7b140c6344/kube-multus/1.log" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.313552 4708 generic.go:334] "Generic (PLEG): container finished" podID="abcfd1aa-ec01-4cae-8858-0d7b140c6344" containerID="78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca" exitCode=2 Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.313651 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerDied","Data":"78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.314911 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"6cfd7878429b77d861de55f6401fdda7fd677eff48ebc6dfa94ec6aa883f4855"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.314957 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"672e0618a41d664b8375415fd94c0799d59202a4ee97a3f1acfa4a3d42de6ec2"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.314978 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"b627d40d9ce12147a78889a1c8b4f39bc51570bdeb4011ec6f009ffcb94ad2d1"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.314982 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.314997 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"98fd94062cfd0f2e7c44204a270ef6b71106349facd758fe6fdcb73ff8a75bde"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315137 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"b82778a04951ec9d2d0f62573a8d483ace202d5361a57de1ff5d2770fbe89ea1"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315155 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"1d36aa314219f36c21102692007536cfe99cb6935d31301bd483ed3c510ac771"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315168 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"59cae9b139cdece3b751c5b6ffba80421f432a9f0a26244d472d5c15d70613b6"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315183 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"f6a98e1b2293c7d650564e41b21b022f5722a466181354d7803980112bcf12bb"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315199 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"62c4fbe35bacbd0f5d700662fad3549c34b9b2cd69ea8ffac257408b1ecc8d88"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315212 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qppcd" event={"ID":"6d78be28-7e93-4fda-98cd-521398f9b3e4","Type":"ContainerDied","Data":"2ac6445092c4cdcbc3bd15dacca5f99b8cc64c35e81d1c36d26b4e7179ff3ff7"} Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.315272 4708 scope.go:117] "RemoveContainer" containerID="78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca" Nov 25 05:48:59 crc kubenswrapper[4708]: E1125 05:48:59.315514 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-rjrnw_openshift-multus(abcfd1aa-ec01-4cae-8858-0d7b140c6344)\"" pod="openshift-multus/multus-rjrnw" podUID="abcfd1aa-ec01-4cae-8858-0d7b140c6344" Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.355223 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qppcd"] Nov 25 05:48:59 crc kubenswrapper[4708]: I1125 05:48:59.362070 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qppcd"] Nov 25 05:49:00 crc kubenswrapper[4708]: I1125 05:49:00.321077 4708 generic.go:334] "Generic (PLEG): container finished" podID="81ff213f-6eda-4128-bee4-8f428c972f61" containerID="fc0de70579ae80c72d65a32fc96de3a596a96b00b85ef87fcdc9b0b8dd6b8121" exitCode=0 Nov 25 05:49:00 crc kubenswrapper[4708]: I1125 05:49:00.321141 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerDied","Data":"fc0de70579ae80c72d65a32fc96de3a596a96b00b85ef87fcdc9b0b8dd6b8121"} Nov 25 05:49:00 crc kubenswrapper[4708]: I1125 05:49:00.899394 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d78be28-7e93-4fda-98cd-521398f9b3e4" path="/var/lib/kubelet/pods/6d78be28-7e93-4fda-98cd-521398f9b3e4/volumes" Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329411 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"08e3fd956a6b886017eb96d67d79771a57d35182ee438033881d363746d0d5fe"} Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329748 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"a494a68b1358d8396f4235d60e36378393422228568a91e1bc96e9ff8c917371"} Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329761 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"8b6deb72a437c6336d7eb0be0e29b2477f3426cf880246bb449e267c6c5efa1f"} Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329771 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"dbbbd58deb3e4162fc0a7ea3205b29dd9b3692ee3e4db834ed6707f7c619c6a4"} Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329780 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"0522b45f9c7301dd2cd3fc0ae0ab5d9d75dd19500919019c429cc651153e5b06"} Nov 25 05:49:01 crc kubenswrapper[4708]: I1125 05:49:01.329791 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"059be9e187dc2a05d62274eba96edbe161c12eda98b3216f8346b1383c1e1f95"} Nov 25 05:49:03 crc kubenswrapper[4708]: I1125 05:49:03.346572 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"90969faf9a576d3df862679796b00ab817c1657b6c4defdf4134732d70c17637"} Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.359911 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" event={"ID":"81ff213f-6eda-4128-bee4-8f428c972f61","Type":"ContainerStarted","Data":"cf036038edaa6bf865a84ce246d592bb637829a8f5f18ce8ca35a4cd917d4131"} Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.360125 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.360158 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.360232 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.381757 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.384750 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:05 crc kubenswrapper[4708]: I1125 05:49:05.388911 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" podStartSLOduration=7.38890167 podStartE2EDuration="7.38890167s" podCreationTimestamp="2025-11-25 05:48:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:49:05.386085321 +0000 UTC m=+486.794918707" watchObservedRunningTime="2025-11-25 05:49:05.38890167 +0000 UTC m=+486.797735056" Nov 25 05:49:12 crc kubenswrapper[4708]: I1125 05:49:12.893017 4708 scope.go:117] "RemoveContainer" containerID="78be18b6bd21fc746123347353e9666f918466d1219df56fac7ecdccf3cdceca" Nov 25 05:49:13 crc kubenswrapper[4708]: I1125 05:49:13.398296 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-rjrnw_abcfd1aa-ec01-4cae-8858-0d7b140c6344/kube-multus/1.log" Nov 25 05:49:13 crc kubenswrapper[4708]: I1125 05:49:13.398593 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-rjrnw" event={"ID":"abcfd1aa-ec01-4cae-8858-0d7b140c6344","Type":"ContainerStarted","Data":"742102bf63140618e958fb68675b0b8d856f10ae392100b450cec61aaebb15ef"} Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.693456 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j"] Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.695027 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.697148 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.704086 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j"] Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.769674 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.769746 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.769916 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvxkm\" (UniqueName: \"kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.870714 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.870779 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvxkm\" (UniqueName: \"kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.870813 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.871224 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.871239 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:23 crc kubenswrapper[4708]: I1125 05:49:23.887631 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvxkm\" (UniqueName: \"kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:24 crc kubenswrapper[4708]: I1125 05:49:24.008398 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:24 crc kubenswrapper[4708]: I1125 05:49:24.350564 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j"] Nov 25 05:49:24 crc kubenswrapper[4708]: W1125 05:49:24.356122 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d46ed7f_a1de_40aa_8b62_f59e5770c6ad.slice/crio-52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28 WatchSource:0}: Error finding container 52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28: Status 404 returned error can't find the container with id 52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28 Nov 25 05:49:24 crc kubenswrapper[4708]: I1125 05:49:24.445463 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" event={"ID":"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad","Type":"ContainerStarted","Data":"52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28"} Nov 25 05:49:25 crc kubenswrapper[4708]: I1125 05:49:25.450866 4708 generic.go:334] "Generic (PLEG): container finished" podID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerID="66019216cf1c104efe6fef23cbb48276fc9ea628ead36d98736fd2e14bf90bba" exitCode=0 Nov 25 05:49:25 crc kubenswrapper[4708]: I1125 05:49:25.450917 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" event={"ID":"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad","Type":"ContainerDied","Data":"66019216cf1c104efe6fef23cbb48276fc9ea628ead36d98736fd2e14bf90bba"} Nov 25 05:49:27 crc kubenswrapper[4708]: I1125 05:49:27.461566 4708 generic.go:334] "Generic (PLEG): container finished" podID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerID="2c389465afa13560aa4e4a60de7d048dc7c5cb2bb95ac24919d2ce6fb8c55a03" exitCode=0 Nov 25 05:49:27 crc kubenswrapper[4708]: I1125 05:49:27.461888 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" event={"ID":"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad","Type":"ContainerDied","Data":"2c389465afa13560aa4e4a60de7d048dc7c5cb2bb95ac24919d2ce6fb8c55a03"} Nov 25 05:49:28 crc kubenswrapper[4708]: I1125 05:49:28.468505 4708 generic.go:334] "Generic (PLEG): container finished" podID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerID="c8eae6c9dadce265de7fa9bc147abbe687fcba036f0973225cee552b4699d4dd" exitCode=0 Nov 25 05:49:28 crc kubenswrapper[4708]: I1125 05:49:28.468587 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" event={"ID":"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad","Type":"ContainerDied","Data":"c8eae6c9dadce265de7fa9bc147abbe687fcba036f0973225cee552b4699d4dd"} Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.281933 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rvg5p" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.670339 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.834582 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvxkm\" (UniqueName: \"kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm\") pod \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.834720 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle\") pod \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.834821 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util\") pod \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\" (UID: \"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad\") " Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.835298 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle" (OuterVolumeSpecName: "bundle") pod "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" (UID: "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.842027 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util" (OuterVolumeSpecName: "util") pod "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" (UID: "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.850604 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm" (OuterVolumeSpecName: "kube-api-access-hvxkm") pod "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" (UID: "0d46ed7f-a1de-40aa-8b62-f59e5770c6ad"). InnerVolumeSpecName "kube-api-access-hvxkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.936050 4708 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-util\") on node \"crc\" DevicePath \"\"" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.936076 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvxkm\" (UniqueName: \"kubernetes.io/projected/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-kube-api-access-hvxkm\") on node \"crc\" DevicePath \"\"" Nov 25 05:49:29 crc kubenswrapper[4708]: I1125 05:49:29.936088 4708 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0d46ed7f-a1de-40aa-8b62-f59e5770c6ad-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:49:30 crc kubenswrapper[4708]: I1125 05:49:30.484188 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" event={"ID":"0d46ed7f-a1de-40aa-8b62-f59e5770c6ad","Type":"ContainerDied","Data":"52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28"} Nov 25 05:49:30 crc kubenswrapper[4708]: I1125 05:49:30.484234 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52538fc31af8e4fd9c5f7e3c1aa9b11545b84bdfaa2a5eb32f509d51c687dd28" Nov 25 05:49:30 crc kubenswrapper[4708]: I1125 05:49:30.484251 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.657927 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-dvth5"] Nov 25 05:49:31 crc kubenswrapper[4708]: E1125 05:49:31.658201 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="extract" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.658215 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="extract" Nov 25 05:49:31 crc kubenswrapper[4708]: E1125 05:49:31.658227 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="util" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.658235 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="util" Nov 25 05:49:31 crc kubenswrapper[4708]: E1125 05:49:31.658259 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="pull" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.658266 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="pull" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.658364 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d46ed7f-a1de-40aa-8b62-f59e5770c6ad" containerName="extract" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.658880 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.660779 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-jhwcl" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.660828 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.661029 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.669366 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-dvth5"] Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.757391 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6dl5\" (UniqueName: \"kubernetes.io/projected/daa36bec-fbbf-44f3-8785-8e09474d7add-kube-api-access-b6dl5\") pod \"nmstate-operator-557fdffb88-dvth5\" (UID: \"daa36bec-fbbf-44f3-8785-8e09474d7add\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.859084 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6dl5\" (UniqueName: \"kubernetes.io/projected/daa36bec-fbbf-44f3-8785-8e09474d7add-kube-api-access-b6dl5\") pod \"nmstate-operator-557fdffb88-dvth5\" (UID: \"daa36bec-fbbf-44f3-8785-8e09474d7add\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.876029 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6dl5\" (UniqueName: \"kubernetes.io/projected/daa36bec-fbbf-44f3-8785-8e09474d7add-kube-api-access-b6dl5\") pod \"nmstate-operator-557fdffb88-dvth5\" (UID: \"daa36bec-fbbf-44f3-8785-8e09474d7add\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" Nov 25 05:49:31 crc kubenswrapper[4708]: I1125 05:49:31.971986 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" Nov 25 05:49:32 crc kubenswrapper[4708]: I1125 05:49:32.363991 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-dvth5"] Nov 25 05:49:32 crc kubenswrapper[4708]: W1125 05:49:32.368592 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaa36bec_fbbf_44f3_8785_8e09474d7add.slice/crio-7a73dcb0882d7787e7a9bb10ffaf0fd7b6b2580af302be561caf966188acbcfa WatchSource:0}: Error finding container 7a73dcb0882d7787e7a9bb10ffaf0fd7b6b2580af302be561caf966188acbcfa: Status 404 returned error can't find the container with id 7a73dcb0882d7787e7a9bb10ffaf0fd7b6b2580af302be561caf966188acbcfa Nov 25 05:49:32 crc kubenswrapper[4708]: I1125 05:49:32.493962 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" event={"ID":"daa36bec-fbbf-44f3-8785-8e09474d7add","Type":"ContainerStarted","Data":"7a73dcb0882d7787e7a9bb10ffaf0fd7b6b2580af302be561caf966188acbcfa"} Nov 25 05:49:34 crc kubenswrapper[4708]: I1125 05:49:34.507541 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" event={"ID":"daa36bec-fbbf-44f3-8785-8e09474d7add","Type":"ContainerStarted","Data":"f343e6dee2f76f03c33ef7a3805644467d859ac7e0ae84eb1fd49ea8728f9698"} Nov 25 05:49:34 crc kubenswrapper[4708]: I1125 05:49:34.533183 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-dvth5" podStartSLOduration=1.656007181 podStartE2EDuration="3.533167837s" podCreationTimestamp="2025-11-25 05:49:31 +0000 UTC" firstStartedPulling="2025-11-25 05:49:32.372026041 +0000 UTC m=+513.780859427" lastFinishedPulling="2025-11-25 05:49:34.249186696 +0000 UTC m=+515.658020083" observedRunningTime="2025-11-25 05:49:34.531086933 +0000 UTC m=+515.939920320" watchObservedRunningTime="2025-11-25 05:49:34.533167837 +0000 UTC m=+515.942001222" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.337924 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.338894 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.340311 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-ljj6h" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.356373 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.359293 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.360046 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.362320 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.368703 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.381622 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-gghd9"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.383892 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.407861 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-nmstate-lock\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.407916 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/03a47fb5-3577-45b2-83b2-cedcd799eea0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.408074 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-dbus-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.408118 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k7cv\" (UniqueName: \"kubernetes.io/projected/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-kube-api-access-7k7cv\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.408161 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlmhb\" (UniqueName: \"kubernetes.io/projected/f7336ae4-7e4b-48a3-8fbd-618e0491a2ff-kube-api-access-jlmhb\") pod \"nmstate-metrics-5dcf9c57c5-kjddh\" (UID: \"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.408178 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-ovs-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.408204 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvdxv\" (UniqueName: \"kubernetes.io/projected/03a47fb5-3577-45b2-83b2-cedcd799eea0-kube-api-access-cvdxv\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.448230 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.449159 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.450838 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.451040 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.453637 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.453981 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-pkkk7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510175 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-dbus-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510273 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k7cv\" (UniqueName: \"kubernetes.io/projected/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-kube-api-access-7k7cv\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510337 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlmhb\" (UniqueName: \"kubernetes.io/projected/f7336ae4-7e4b-48a3-8fbd-618e0491a2ff-kube-api-access-jlmhb\") pod \"nmstate-metrics-5dcf9c57c5-kjddh\" (UID: \"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510367 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-ovs-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510405 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvdxv\" (UniqueName: \"kubernetes.io/projected/03a47fb5-3577-45b2-83b2-cedcd799eea0-kube-api-access-cvdxv\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510434 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-nmstate-lock\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.510486 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/03a47fb5-3577-45b2-83b2-cedcd799eea0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.511149 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-dbus-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.511590 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-nmstate-lock\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.511645 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-ovs-socket\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.528102 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k7cv\" (UniqueName: \"kubernetes.io/projected/7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7-kube-api-access-7k7cv\") pod \"nmstate-handler-gghd9\" (UID: \"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7\") " pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.529031 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlmhb\" (UniqueName: \"kubernetes.io/projected/f7336ae4-7e4b-48a3-8fbd-618e0491a2ff-kube-api-access-jlmhb\") pod \"nmstate-metrics-5dcf9c57c5-kjddh\" (UID: \"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.530410 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/03a47fb5-3577-45b2-83b2-cedcd799eea0-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.531036 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvdxv\" (UniqueName: \"kubernetes.io/projected/03a47fb5-3577-45b2-83b2-cedcd799eea0-kube-api-access-cvdxv\") pod \"nmstate-webhook-6b89b748d8-2xlzf\" (UID: \"03a47fb5-3577-45b2-83b2-cedcd799eea0\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.613368 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6725355-22e6-4236-94e9-ce851f0d8e29-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.613708 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d6725355-22e6-4236-94e9-ce851f0d8e29-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.613732 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2rj4\" (UniqueName: \"kubernetes.io/projected/d6725355-22e6-4236-94e9-ce851f0d8e29-kube-api-access-c2rj4\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.620374 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7ddf657454-dp9d7"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.621237 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.633503 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ddf657454-dp9d7"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.651591 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.673175 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.694902 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:35 crc kubenswrapper[4708]: W1125 05:49:35.710617 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7590e799_a5e1_4fc1_b9ba_c3d6d9084cb7.slice/crio-59922380c51b12f7308d4f3aa15560881f890875e6a3544a6681f66978e90bbc WatchSource:0}: Error finding container 59922380c51b12f7308d4f3aa15560881f890875e6a3544a6681f66978e90bbc: Status 404 returned error can't find the container with id 59922380c51b12f7308d4f3aa15560881f890875e6a3544a6681f66978e90bbc Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.714885 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d6725355-22e6-4236-94e9-ce851f0d8e29-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.715194 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2rj4\" (UniqueName: \"kubernetes.io/projected/d6725355-22e6-4236-94e9-ce851f0d8e29-kube-api-access-c2rj4\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.715328 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6725355-22e6-4236-94e9-ce851f0d8e29-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.716033 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d6725355-22e6-4236-94e9-ce851f0d8e29-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.721746 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d6725355-22e6-4236-94e9-ce851f0d8e29-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.734881 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2rj4\" (UniqueName: \"kubernetes.io/projected/d6725355-22e6-4236-94e9-ce851f0d8e29-kube-api-access-c2rj4\") pod \"nmstate-console-plugin-5874bd7bc5-w982m\" (UID: \"d6725355-22e6-4236-94e9-ce851f0d8e29\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.768110 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818128 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-service-ca\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818179 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818310 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-oauth-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818352 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-oauth-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818395 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818441 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-trusted-ca-bundle\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.818470 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rgl4\" (UniqueName: \"kubernetes.io/projected/010be2cb-c932-49d4-8277-f2ab07b8dcde-kube-api-access-9rgl4\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.882788 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf"] Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920080 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rgl4\" (UniqueName: \"kubernetes.io/projected/010be2cb-c932-49d4-8277-f2ab07b8dcde-kube-api-access-9rgl4\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920170 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-service-ca\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920205 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920252 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-oauth-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920278 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-oauth-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920303 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.920335 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-trusted-ca-bundle\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.921637 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-oauth-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.921790 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.922844 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-trusted-ca-bundle\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.922945 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/010be2cb-c932-49d4-8277-f2ab07b8dcde-service-ca\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.925664 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-serving-cert\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.928860 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/010be2cb-c932-49d4-8277-f2ab07b8dcde-console-oauth-config\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:35 crc kubenswrapper[4708]: I1125 05:49:35.935983 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rgl4\" (UniqueName: \"kubernetes.io/projected/010be2cb-c932-49d4-8277-f2ab07b8dcde-kube-api-access-9rgl4\") pod \"console-7ddf657454-dp9d7\" (UID: \"010be2cb-c932-49d4-8277-f2ab07b8dcde\") " pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.023567 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh"] Nov 25 05:49:36 crc kubenswrapper[4708]: W1125 05:49:36.026179 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7336ae4_7e4b_48a3_8fbd_618e0491a2ff.slice/crio-09ba0c904aaddf11a1f34696da21d70a87d58450fd7129102495f17e87f73dfc WatchSource:0}: Error finding container 09ba0c904aaddf11a1f34696da21d70a87d58450fd7129102495f17e87f73dfc: Status 404 returned error can't find the container with id 09ba0c904aaddf11a1f34696da21d70a87d58450fd7129102495f17e87f73dfc Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.151164 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m"] Nov 25 05:49:36 crc kubenswrapper[4708]: W1125 05:49:36.154477 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6725355_22e6_4236_94e9_ce851f0d8e29.slice/crio-51d9903b7ac9ac56d65a180c24e6f30834b4703cf74c9dcaa877d63488991aaf WatchSource:0}: Error finding container 51d9903b7ac9ac56d65a180c24e6f30834b4703cf74c9dcaa877d63488991aaf: Status 404 returned error can't find the container with id 51d9903b7ac9ac56d65a180c24e6f30834b4703cf74c9dcaa877d63488991aaf Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.235206 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.520423 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" event={"ID":"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff","Type":"ContainerStarted","Data":"09ba0c904aaddf11a1f34696da21d70a87d58450fd7129102495f17e87f73dfc"} Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.521821 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gghd9" event={"ID":"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7","Type":"ContainerStarted","Data":"59922380c51b12f7308d4f3aa15560881f890875e6a3544a6681f66978e90bbc"} Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.523841 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" event={"ID":"d6725355-22e6-4236-94e9-ce851f0d8e29","Type":"ContainerStarted","Data":"51d9903b7ac9ac56d65a180c24e6f30834b4703cf74c9dcaa877d63488991aaf"} Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.524826 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" event={"ID":"03a47fb5-3577-45b2-83b2-cedcd799eea0","Type":"ContainerStarted","Data":"0b3d92419de8d1b6e16f77d4548ca7890032b6998cdeae4aed8b873ba588ebac"} Nov 25 05:49:36 crc kubenswrapper[4708]: I1125 05:49:36.610790 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7ddf657454-dp9d7"] Nov 25 05:49:36 crc kubenswrapper[4708]: W1125 05:49:36.615387 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod010be2cb_c932_49d4_8277_f2ab07b8dcde.slice/crio-351cb31ddc3e041ec028e30bf306289a25cb1bb7afc7e1a4e3dd151a5ccde874 WatchSource:0}: Error finding container 351cb31ddc3e041ec028e30bf306289a25cb1bb7afc7e1a4e3dd151a5ccde874: Status 404 returned error can't find the container with id 351cb31ddc3e041ec028e30bf306289a25cb1bb7afc7e1a4e3dd151a5ccde874 Nov 25 05:49:37 crc kubenswrapper[4708]: I1125 05:49:37.534758 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ddf657454-dp9d7" event={"ID":"010be2cb-c932-49d4-8277-f2ab07b8dcde","Type":"ContainerStarted","Data":"8104c112f302f500125ce7cde1339374e4568ed0134edb22eff70fcfe40cedb4"} Nov 25 05:49:37 crc kubenswrapper[4708]: I1125 05:49:37.535156 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7ddf657454-dp9d7" event={"ID":"010be2cb-c932-49d4-8277-f2ab07b8dcde","Type":"ContainerStarted","Data":"351cb31ddc3e041ec028e30bf306289a25cb1bb7afc7e1a4e3dd151a5ccde874"} Nov 25 05:49:37 crc kubenswrapper[4708]: I1125 05:49:37.559689 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7ddf657454-dp9d7" podStartSLOduration=2.559649746 podStartE2EDuration="2.559649746s" podCreationTimestamp="2025-11-25 05:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:49:37.555120668 +0000 UTC m=+518.963954054" watchObservedRunningTime="2025-11-25 05:49:37.559649746 +0000 UTC m=+518.968483132" Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.552829 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" event={"ID":"03a47fb5-3577-45b2-83b2-cedcd799eea0","Type":"ContainerStarted","Data":"affaced7f227eeb3d23b85248bab4d2f7270f3a6729d6378fdc2ae33cf5b1f39"} Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.553066 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.554085 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" event={"ID":"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff","Type":"ContainerStarted","Data":"c32448f0ca527050c2c8d8f2756e5412b67d1f88dddbc4004b3946b6eed4f2d6"} Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.555477 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-gghd9" event={"ID":"7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7","Type":"ContainerStarted","Data":"8a7d1d8e8ed35aa699985edc84912eed17c9882808b05e918d56af503c4a159e"} Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.555573 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.557635 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" event={"ID":"d6725355-22e6-4236-94e9-ce851f0d8e29","Type":"ContainerStarted","Data":"99ec0b69ca693bc6847fee586602af890e01c1e42f25d6143398d77948f970f5"} Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.571905 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" podStartSLOduration=1.204577909 podStartE2EDuration="4.571887713s" podCreationTimestamp="2025-11-25 05:49:35 +0000 UTC" firstStartedPulling="2025-11-25 05:49:35.895746976 +0000 UTC m=+517.304580362" lastFinishedPulling="2025-11-25 05:49:39.26305678 +0000 UTC m=+520.671890166" observedRunningTime="2025-11-25 05:49:39.565291401 +0000 UTC m=+520.974124797" watchObservedRunningTime="2025-11-25 05:49:39.571887713 +0000 UTC m=+520.980721099" Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.581266 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-w982m" podStartSLOduration=1.479030603 podStartE2EDuration="4.581253637s" podCreationTimestamp="2025-11-25 05:49:35 +0000 UTC" firstStartedPulling="2025-11-25 05:49:36.157202364 +0000 UTC m=+517.566035750" lastFinishedPulling="2025-11-25 05:49:39.259425398 +0000 UTC m=+520.668258784" observedRunningTime="2025-11-25 05:49:39.574784553 +0000 UTC m=+520.983617940" watchObservedRunningTime="2025-11-25 05:49:39.581253637 +0000 UTC m=+520.990087023" Nov 25 05:49:39 crc kubenswrapper[4708]: I1125 05:49:39.589351 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-gghd9" podStartSLOduration=1.042723538 podStartE2EDuration="4.589332378s" podCreationTimestamp="2025-11-25 05:49:35 +0000 UTC" firstStartedPulling="2025-11-25 05:49:35.712634445 +0000 UTC m=+517.121467831" lastFinishedPulling="2025-11-25 05:49:39.259243284 +0000 UTC m=+520.668076671" observedRunningTime="2025-11-25 05:49:39.584930465 +0000 UTC m=+520.993763841" watchObservedRunningTime="2025-11-25 05:49:39.589332378 +0000 UTC m=+520.998165764" Nov 25 05:49:42 crc kubenswrapper[4708]: I1125 05:49:42.577778 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" event={"ID":"f7336ae4-7e4b-48a3-8fbd-618e0491a2ff","Type":"ContainerStarted","Data":"cb723ee147ae16e705b4f37b1b2f0be1afdb2cca001bf7de0da13e8ae7abbde7"} Nov 25 05:49:42 crc kubenswrapper[4708]: I1125 05:49:42.593289 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-kjddh" podStartSLOduration=1.949006434 podStartE2EDuration="7.593272618s" podCreationTimestamp="2025-11-25 05:49:35 +0000 UTC" firstStartedPulling="2025-11-25 05:49:36.029767236 +0000 UTC m=+517.438600622" lastFinishedPulling="2025-11-25 05:49:41.67403342 +0000 UTC m=+523.082866806" observedRunningTime="2025-11-25 05:49:42.592243281 +0000 UTC m=+524.001076667" watchObservedRunningTime="2025-11-25 05:49:42.593272618 +0000 UTC m=+524.002106004" Nov 25 05:49:44 crc kubenswrapper[4708]: I1125 05:49:44.356914 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:49:44 crc kubenswrapper[4708]: I1125 05:49:44.357959 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:49:45 crc kubenswrapper[4708]: I1125 05:49:45.717441 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-gghd9" Nov 25 05:49:46 crc kubenswrapper[4708]: I1125 05:49:46.235488 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:46 crc kubenswrapper[4708]: I1125 05:49:46.235631 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:46 crc kubenswrapper[4708]: I1125 05:49:46.241083 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:46 crc kubenswrapper[4708]: I1125 05:49:46.603034 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7ddf657454-dp9d7" Nov 25 05:49:46 crc kubenswrapper[4708]: I1125 05:49:46.645060 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:49:55 crc kubenswrapper[4708]: I1125 05:49:55.680982 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-2xlzf" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.851362 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn"] Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.853865 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.859042 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.859574 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn"] Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.884281 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.884341 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5txw\" (UniqueName: \"kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.884383 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.985911 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.985978 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5txw\" (UniqueName: \"kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.986023 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.986606 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:05 crc kubenswrapper[4708]: I1125 05:50:05.986689 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:06 crc kubenswrapper[4708]: I1125 05:50:06.004641 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5txw\" (UniqueName: \"kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:06 crc kubenswrapper[4708]: I1125 05:50:06.169807 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:06 crc kubenswrapper[4708]: I1125 05:50:06.531733 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn"] Nov 25 05:50:06 crc kubenswrapper[4708]: I1125 05:50:06.719954 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerStarted","Data":"50fb2a1b9b95a09ce059815339b55b3c09699dcff1f1551f030f5de48e8b9819"} Nov 25 05:50:06 crc kubenswrapper[4708]: I1125 05:50:06.720305 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerStarted","Data":"ac4c7061d8235bb575865b7b3218cd78a5408ea2b5aaa78fc5d83144cc93adb6"} Nov 25 05:50:07 crc kubenswrapper[4708]: I1125 05:50:07.729708 4708 generic.go:334] "Generic (PLEG): container finished" podID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerID="50fb2a1b9b95a09ce059815339b55b3c09699dcff1f1551f030f5de48e8b9819" exitCode=0 Nov 25 05:50:07 crc kubenswrapper[4708]: I1125 05:50:07.729771 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerDied","Data":"50fb2a1b9b95a09ce059815339b55b3c09699dcff1f1551f030f5de48e8b9819"} Nov 25 05:50:09 crc kubenswrapper[4708]: I1125 05:50:09.744113 4708 generic.go:334] "Generic (PLEG): container finished" podID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerID="f23187d19c159d538955f919baab97a686be0c3f5f7ebdce0bc591082fb7201e" exitCode=0 Nov 25 05:50:09 crc kubenswrapper[4708]: I1125 05:50:09.744244 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerDied","Data":"f23187d19c159d538955f919baab97a686be0c3f5f7ebdce0bc591082fb7201e"} Nov 25 05:50:10 crc kubenswrapper[4708]: I1125 05:50:10.753499 4708 generic.go:334] "Generic (PLEG): container finished" podID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerID="1b67c65a6fe713dea43c7b0b43509bd7dd6c16d2909ea11abbcbab86c8faf345" exitCode=0 Nov 25 05:50:10 crc kubenswrapper[4708]: I1125 05:50:10.753551 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerDied","Data":"1b67c65a6fe713dea43c7b0b43509bd7dd6c16d2909ea11abbcbab86c8faf345"} Nov 25 05:50:11 crc kubenswrapper[4708]: I1125 05:50:11.679317 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zkmtb" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerName="console" containerID="cri-o://76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45" gracePeriod=15 Nov 25 05:50:11 crc kubenswrapper[4708]: I1125 05:50:11.991231 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:11 crc kubenswrapper[4708]: I1125 05:50:11.995837 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkmtb_c3776f93-7ce9-4056-86c8-7a2d95a02650/console/0.log" Nov 25 05:50:11 crc kubenswrapper[4708]: I1125 05:50:11.995968 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.056902 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.056940 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.056964 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057009 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle\") pod \"615985f2-6094-4f12-8ab8-c19cb9d18971\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057041 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q5txw\" (UniqueName: \"kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw\") pod \"615985f2-6094-4f12-8ab8-c19cb9d18971\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057062 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgs6c\" (UniqueName: \"kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057088 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057112 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057140 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert\") pod \"c3776f93-7ce9-4056-86c8-7a2d95a02650\" (UID: \"c3776f93-7ce9-4056-86c8-7a2d95a02650\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.057189 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util\") pod \"615985f2-6094-4f12-8ab8-c19cb9d18971\" (UID: \"615985f2-6094-4f12-8ab8-c19cb9d18971\") " Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.058087 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config" (OuterVolumeSpecName: "console-config") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.058114 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.058208 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca" (OuterVolumeSpecName: "service-ca") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.058588 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.058644 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle" (OuterVolumeSpecName: "bundle") pod "615985f2-6094-4f12-8ab8-c19cb9d18971" (UID: "615985f2-6094-4f12-8ab8-c19cb9d18971"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.063012 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw" (OuterVolumeSpecName: "kube-api-access-q5txw") pod "615985f2-6094-4f12-8ab8-c19cb9d18971" (UID: "615985f2-6094-4f12-8ab8-c19cb9d18971"). InnerVolumeSpecName "kube-api-access-q5txw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.063420 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.064100 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c" (OuterVolumeSpecName: "kube-api-access-pgs6c") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "kube-api-access-pgs6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.064141 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "c3776f93-7ce9-4056-86c8-7a2d95a02650" (UID: "c3776f93-7ce9-4056-86c8-7a2d95a02650"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.158403 4708 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.158577 4708 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.158689 4708 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.158954 4708 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.159032 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q5txw\" (UniqueName: \"kubernetes.io/projected/615985f2-6094-4f12-8ab8-c19cb9d18971-kube-api-access-q5txw\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.159107 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgs6c\" (UniqueName: \"kubernetes.io/projected/c3776f93-7ce9-4056-86c8-7a2d95a02650-kube-api-access-pgs6c\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.159180 4708 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/c3776f93-7ce9-4056-86c8-7a2d95a02650-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.159232 4708 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.159307 4708 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/c3776f93-7ce9-4056-86c8-7a2d95a02650-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.241227 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util" (OuterVolumeSpecName: "util") pod "615985f2-6094-4f12-8ab8-c19cb9d18971" (UID: "615985f2-6094-4f12-8ab8-c19cb9d18971"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.260097 4708 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/615985f2-6094-4f12-8ab8-c19cb9d18971-util\") on node \"crc\" DevicePath \"\"" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.765709 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zkmtb_c3776f93-7ce9-4056-86c8-7a2d95a02650/console/0.log" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.765756 4708 generic.go:334] "Generic (PLEG): container finished" podID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerID="76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45" exitCode=2 Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.765825 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zkmtb" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.765835 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkmtb" event={"ID":"c3776f93-7ce9-4056-86c8-7a2d95a02650","Type":"ContainerDied","Data":"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45"} Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.766008 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zkmtb" event={"ID":"c3776f93-7ce9-4056-86c8-7a2d95a02650","Type":"ContainerDied","Data":"030fd166ce25a6d774455f427f1f904879725265278f754454cbe09f4f876fa7"} Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.766070 4708 scope.go:117] "RemoveContainer" containerID="76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.768103 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" event={"ID":"615985f2-6094-4f12-8ab8-c19cb9d18971","Type":"ContainerDied","Data":"ac4c7061d8235bb575865b7b3218cd78a5408ea2b5aaa78fc5d83144cc93adb6"} Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.768193 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac4c7061d8235bb575865b7b3218cd78a5408ea2b5aaa78fc5d83144cc93adb6" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.768301 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.781368 4708 scope.go:117] "RemoveContainer" containerID="76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45" Nov 25 05:50:12 crc kubenswrapper[4708]: E1125 05:50:12.782147 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45\": container with ID starting with 76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45 not found: ID does not exist" containerID="76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.782176 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45"} err="failed to get container status \"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45\": rpc error: code = NotFound desc = could not find container \"76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45\": container with ID starting with 76a9744d6af2ce9cb303a9508e501c7ea06acff3a1c8971f310f516073dd6a45 not found: ID does not exist" Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.790179 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.792656 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zkmtb"] Nov 25 05:50:12 crc kubenswrapper[4708]: I1125 05:50:12.900511 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" path="/var/lib/kubelet/pods/c3776f93-7ce9-4056-86c8-7a2d95a02650/volumes" Nov 25 05:50:14 crc kubenswrapper[4708]: I1125 05:50:14.356973 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:50:14 crc kubenswrapper[4708]: I1125 05:50:14.357335 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.877368 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz"] Nov 25 05:50:21 crc kubenswrapper[4708]: E1125 05:50:21.877965 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerName="console" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.877981 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerName="console" Nov 25 05:50:21 crc kubenswrapper[4708]: E1125 05:50:21.878002 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="util" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878007 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="util" Nov 25 05:50:21 crc kubenswrapper[4708]: E1125 05:50:21.878017 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="pull" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878023 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="pull" Nov 25 05:50:21 crc kubenswrapper[4708]: E1125 05:50:21.878033 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="extract" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878039 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="extract" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878142 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3776f93-7ce9-4056-86c8-7a2d95a02650" containerName="console" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878153 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="615985f2-6094-4f12-8ab8-c19cb9d18971" containerName="extract" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.878625 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.882756 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.882802 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.882878 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.883736 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.885629 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-fdt9w" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.895013 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz"] Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.982899 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vbpj\" (UniqueName: \"kubernetes.io/projected/93f12b13-309e-4a76-9f10-c1ca9a0545eb-kube-api-access-5vbpj\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.982977 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-apiservice-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:21 crc kubenswrapper[4708]: I1125 05:50:21.983233 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-webhook-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.084648 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-apiservice-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.085028 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-webhook-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.085135 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vbpj\" (UniqueName: \"kubernetes.io/projected/93f12b13-309e-4a76-9f10-c1ca9a0545eb-kube-api-access-5vbpj\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.097140 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-webhook-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.105390 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/93f12b13-309e-4a76-9f10-c1ca9a0545eb-apiservice-cert\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.125178 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx"] Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.125970 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.128191 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-ccln5" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.128386 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.129463 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.133117 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vbpj\" (UniqueName: \"kubernetes.io/projected/93f12b13-309e-4a76-9f10-c1ca9a0545eb-kube-api-access-5vbpj\") pod \"metallb-operator-controller-manager-5885d8f89f-gjghz\" (UID: \"93f12b13-309e-4a76-9f10-c1ca9a0545eb\") " pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.156527 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx"] Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.197776 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.291560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-webhook-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.291627 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vr7j\" (UniqueName: \"kubernetes.io/projected/9ae15378-5b01-442c-a1b3-52d25e474875-kube-api-access-2vr7j\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.291666 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-apiservice-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.393676 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vr7j\" (UniqueName: \"kubernetes.io/projected/9ae15378-5b01-442c-a1b3-52d25e474875-kube-api-access-2vr7j\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.393799 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-apiservice-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.394649 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-webhook-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.400286 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-apiservice-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.400304 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9ae15378-5b01-442c-a1b3-52d25e474875-webhook-cert\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.412484 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vr7j\" (UniqueName: \"kubernetes.io/projected/9ae15378-5b01-442c-a1b3-52d25e474875-kube-api-access-2vr7j\") pod \"metallb-operator-webhook-server-66489dff8c-hbstx\" (UID: \"9ae15378-5b01-442c-a1b3-52d25e474875\") " pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.454465 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.615224 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz"] Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.809240 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx"] Nov 25 05:50:22 crc kubenswrapper[4708]: W1125 05:50:22.813171 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ae15378_5b01_442c_a1b3_52d25e474875.slice/crio-ab147db6f9b229602a515a2c28554382287725a59b4f95730d80029fb5a62c05 WatchSource:0}: Error finding container ab147db6f9b229602a515a2c28554382287725a59b4f95730d80029fb5a62c05: Status 404 returned error can't find the container with id ab147db6f9b229602a515a2c28554382287725a59b4f95730d80029fb5a62c05 Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.820892 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" event={"ID":"93f12b13-309e-4a76-9f10-c1ca9a0545eb","Type":"ContainerStarted","Data":"c096b85ce0058fe97c76901d120439e2cf9d512655f25f848c58542cf40a5aaf"} Nov 25 05:50:22 crc kubenswrapper[4708]: I1125 05:50:22.822129 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" event={"ID":"9ae15378-5b01-442c-a1b3-52d25e474875","Type":"ContainerStarted","Data":"ab147db6f9b229602a515a2c28554382287725a59b4f95730d80029fb5a62c05"} Nov 25 05:50:26 crc kubenswrapper[4708]: I1125 05:50:26.861331 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" event={"ID":"93f12b13-309e-4a76-9f10-c1ca9a0545eb","Type":"ContainerStarted","Data":"da66b4f3edc9d0e716d474a4d0dc54ae73636cd17f674b099dad2ca7c54a9366"} Nov 25 05:50:26 crc kubenswrapper[4708]: I1125 05:50:26.862140 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:50:26 crc kubenswrapper[4708]: I1125 05:50:26.878969 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" podStartSLOduration=1.939362546 podStartE2EDuration="5.878952904s" podCreationTimestamp="2025-11-25 05:50:21 +0000 UTC" firstStartedPulling="2025-11-25 05:50:22.621350338 +0000 UTC m=+564.030183724" lastFinishedPulling="2025-11-25 05:50:26.560940696 +0000 UTC m=+567.969774082" observedRunningTime="2025-11-25 05:50:26.876950757 +0000 UTC m=+568.285784143" watchObservedRunningTime="2025-11-25 05:50:26.878952904 +0000 UTC m=+568.287786290" Nov 25 05:50:28 crc kubenswrapper[4708]: I1125 05:50:28.872682 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" event={"ID":"9ae15378-5b01-442c-a1b3-52d25e474875","Type":"ContainerStarted","Data":"75a403f004fa12da59b5318814cc75acb21150fa2829039bc0400e3207e99fb2"} Nov 25 05:50:28 crc kubenswrapper[4708]: I1125 05:50:28.872928 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:28 crc kubenswrapper[4708]: I1125 05:50:28.892966 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" podStartSLOduration=1.511549512 podStartE2EDuration="6.892942972s" podCreationTimestamp="2025-11-25 05:50:22 +0000 UTC" firstStartedPulling="2025-11-25 05:50:22.816044664 +0000 UTC m=+564.224878051" lastFinishedPulling="2025-11-25 05:50:28.197438125 +0000 UTC m=+569.606271511" observedRunningTime="2025-11-25 05:50:28.887214284 +0000 UTC m=+570.296047670" watchObservedRunningTime="2025-11-25 05:50:28.892942972 +0000 UTC m=+570.301776359" Nov 25 05:50:42 crc kubenswrapper[4708]: I1125 05:50:42.458782 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-66489dff8c-hbstx" Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.358158 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.358313 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.358643 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.359948 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.360054 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f" gracePeriod=600 Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.971768 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f" exitCode=0 Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.971828 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f"} Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.972107 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977"} Nov 25 05:50:44 crc kubenswrapper[4708]: I1125 05:50:44.972130 4708 scope.go:117] "RemoveContainer" containerID="1cd334abee4d9c754a073c2729219e9b99ccebf128dba5008a0ca8ae2e195dd2" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.200806 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5885d8f89f-gjghz" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.756337 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-f8lcr"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.762241 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.770997 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.771762 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vschs" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.780128 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.783551 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.789791 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.792099 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.793681 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837515 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/415c292a-1706-4c7b-9f9e-30f469708872-frr-startup\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837598 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsdpx\" (UniqueName: \"kubernetes.io/projected/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-kube-api-access-jsdpx\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837618 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/415c292a-1706-4c7b-9f9e-30f469708872-metrics-certs\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837652 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-reloader\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837664 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-conf\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837679 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsmqq\" (UniqueName: \"kubernetes.io/projected/415c292a-1706-4c7b-9f9e-30f469708872-kube-api-access-lsmqq\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837701 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-cert\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837733 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-metrics\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.837765 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-sockets\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.839470 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-dg7zk"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.840559 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dg7zk" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.845115 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.845144 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.845366 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.845706 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-kpnl2" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.849453 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-bkbdq"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.851592 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.855217 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.860235 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-bkbdq"] Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939065 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-reloader\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939102 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-conf\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939122 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsmqq\" (UniqueName: \"kubernetes.io/projected/415c292a-1706-4c7b-9f9e-30f469708872-kube-api-access-lsmqq\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939153 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-cert\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939174 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-cert\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939190 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fb80d906-e252-4e84-95a3-8c63ef010175-metallb-excludel2\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939236 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-metrics\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939260 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c66zj\" (UniqueName: \"kubernetes.io/projected/fb80d906-e252-4e84-95a3-8c63ef010175-kube-api-access-c66zj\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939286 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-sockets\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939300 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-metrics-certs\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939345 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939364 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-metrics-certs\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939384 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/415c292a-1706-4c7b-9f9e-30f469708872-frr-startup\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939416 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsdpx\" (UniqueName: \"kubernetes.io/projected/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-kube-api-access-jsdpx\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939433 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/415c292a-1706-4c7b-9f9e-30f469708872-metrics-certs\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.939448 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx5b7\" (UniqueName: \"kubernetes.io/projected/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-kube-api-access-lx5b7\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.940160 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-conf\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.940232 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-reloader\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.940244 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-metrics\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.940329 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/415c292a-1706-4c7b-9f9e-30f469708872-frr-sockets\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.940982 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/415c292a-1706-4c7b-9f9e-30f469708872-frr-startup\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.949024 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/415c292a-1706-4c7b-9f9e-30f469708872-metrics-certs\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.960199 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-cert\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.963242 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsdpx\" (UniqueName: \"kubernetes.io/projected/f3e32da4-a772-4f5b-b91a-fbd4f5405a94-kube-api-access-jsdpx\") pod \"frr-k8s-webhook-server-6998585d5-5fbpc\" (UID: \"f3e32da4-a772-4f5b-b91a-fbd4f5405a94\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:02 crc kubenswrapper[4708]: I1125 05:51:02.966988 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsmqq\" (UniqueName: \"kubernetes.io/projected/415c292a-1706-4c7b-9f9e-30f469708872-kube-api-access-lsmqq\") pod \"frr-k8s-f8lcr\" (UID: \"415c292a-1706-4c7b-9f9e-30f469708872\") " pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040561 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-metrics-certs\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040635 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx5b7\" (UniqueName: \"kubernetes.io/projected/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-kube-api-access-lx5b7\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040704 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fb80d906-e252-4e84-95a3-8c63ef010175-metallb-excludel2\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040721 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-cert\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: E1125 05:51:03.040744 4708 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040778 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c66zj\" (UniqueName: \"kubernetes.io/projected/fb80d906-e252-4e84-95a3-8c63ef010175-kube-api-access-c66zj\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: E1125 05:51:03.040811 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist podName:fb80d906-e252-4e84-95a3-8c63ef010175 nodeName:}" failed. No retries permitted until 2025-11-25 05:51:03.540792418 +0000 UTC m=+604.949625804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist") pod "speaker-dg7zk" (UID: "fb80d906-e252-4e84-95a3-8c63ef010175") : secret "metallb-memberlist" not found Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.040843 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-metrics-certs\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.041815 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/fb80d906-e252-4e84-95a3-8c63ef010175-metallb-excludel2\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.044919 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-metrics-certs\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.044968 4708 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.045015 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-metrics-certs\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.054137 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-cert\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.055923 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c66zj\" (UniqueName: \"kubernetes.io/projected/fb80d906-e252-4e84-95a3-8c63ef010175-kube-api-access-c66zj\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.058335 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx5b7\" (UniqueName: \"kubernetes.io/projected/56eb3e20-e964-4e88-a2f7-8f07c3e8d40f-kube-api-access-lx5b7\") pod \"controller-6c7b4b5f48-bkbdq\" (UID: \"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f\") " pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.087977 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.104141 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.170057 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.354084 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-bkbdq"] Nov 25 05:51:03 crc kubenswrapper[4708]: W1125 05:51:03.355804 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56eb3e20_e964_4e88_a2f7_8f07c3e8d40f.slice/crio-1cd440a940560bd2f4febf8c875b91b0e964504cd18648ec5097caff6e931f5a WatchSource:0}: Error finding container 1cd440a940560bd2f4febf8c875b91b0e964504cd18648ec5097caff6e931f5a: Status 404 returned error can't find the container with id 1cd440a940560bd2f4febf8c875b91b0e964504cd18648ec5097caff6e931f5a Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.488888 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc"] Nov 25 05:51:03 crc kubenswrapper[4708]: W1125 05:51:03.500548 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3e32da4_a772_4f5b_b91a_fbd4f5405a94.slice/crio-a8b06880f3c39da27dc5c9b4ac6bef1507fe2ef950275c73999893d01aed0c26 WatchSource:0}: Error finding container a8b06880f3c39da27dc5c9b4ac6bef1507fe2ef950275c73999893d01aed0c26: Status 404 returned error can't find the container with id a8b06880f3c39da27dc5c9b4ac6bef1507fe2ef950275c73999893d01aed0c26 Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.551165 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.557247 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/fb80d906-e252-4e84-95a3-8c63ef010175-memberlist\") pod \"speaker-dg7zk\" (UID: \"fb80d906-e252-4e84-95a3-8c63ef010175\") " pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: I1125 05:51:03.756118 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-dg7zk" Nov 25 05:51:03 crc kubenswrapper[4708]: W1125 05:51:03.777463 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb80d906_e252_4e84_95a3_8c63ef010175.slice/crio-fe13be82d9262a98d258e9d6cdb6fecf277ebf280a4ca267d4c9a8b98944eeff WatchSource:0}: Error finding container fe13be82d9262a98d258e9d6cdb6fecf277ebf280a4ca267d4c9a8b98944eeff: Status 404 returned error can't find the container with id fe13be82d9262a98d258e9d6cdb6fecf277ebf280a4ca267d4c9a8b98944eeff Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.099861 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bkbdq" event={"ID":"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f","Type":"ContainerStarted","Data":"8c10ac00d4dcfe48e70065892aa9cf8b07945905adb802a064739383e8bbd3f0"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.100226 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bkbdq" event={"ID":"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f","Type":"ContainerStarted","Data":"59e6c656e8fc558580d08911b42a974bad7a41a660c0affa802fc87fb08e4dcd"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.100240 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bkbdq" event={"ID":"56eb3e20-e964-4e88-a2f7-8f07c3e8d40f","Type":"ContainerStarted","Data":"1cd440a940560bd2f4febf8c875b91b0e964504cd18648ec5097caff6e931f5a"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.100262 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.105618 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dg7zk" event={"ID":"fb80d906-e252-4e84-95a3-8c63ef010175","Type":"ContainerStarted","Data":"4120b5171e35b9f66ad1658af04072d213501e0af5cf8fddf65f33c0bfe3a06c"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.105654 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dg7zk" event={"ID":"fb80d906-e252-4e84-95a3-8c63ef010175","Type":"ContainerStarted","Data":"fe13be82d9262a98d258e9d6cdb6fecf277ebf280a4ca267d4c9a8b98944eeff"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.107393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" event={"ID":"f3e32da4-a772-4f5b-b91a-fbd4f5405a94","Type":"ContainerStarted","Data":"a8b06880f3c39da27dc5c9b4ac6bef1507fe2ef950275c73999893d01aed0c26"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.108824 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"7b51ff5e8cc2d58006a132c3cb2709c701da6d2f34c296b95553f40fe72c8e21"} Nov 25 05:51:04 crc kubenswrapper[4708]: I1125 05:51:04.120147 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-bkbdq" podStartSLOduration=2.120134397 podStartE2EDuration="2.120134397s" podCreationTimestamp="2025-11-25 05:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:51:04.116673616 +0000 UTC m=+605.525507001" watchObservedRunningTime="2025-11-25 05:51:04.120134397 +0000 UTC m=+605.528967783" Nov 25 05:51:05 crc kubenswrapper[4708]: I1125 05:51:05.120497 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-dg7zk" event={"ID":"fb80d906-e252-4e84-95a3-8c63ef010175","Type":"ContainerStarted","Data":"42ebfc4be90f58c7bcc504a407e22182a8150d09c19810fc7f1437c019ec5529"} Nov 25 05:51:05 crc kubenswrapper[4708]: I1125 05:51:05.120969 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-dg7zk" Nov 25 05:51:05 crc kubenswrapper[4708]: I1125 05:51:05.152844 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-dg7zk" podStartSLOduration=3.152817769 podStartE2EDuration="3.152817769s" podCreationTimestamp="2025-11-25 05:51:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:51:05.146816711 +0000 UTC m=+606.555650097" watchObservedRunningTime="2025-11-25 05:51:05.152817769 +0000 UTC m=+606.561651154" Nov 25 05:51:10 crc kubenswrapper[4708]: I1125 05:51:10.155727 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" event={"ID":"f3e32da4-a772-4f5b-b91a-fbd4f5405a94","Type":"ContainerStarted","Data":"2a9e623f18337c45081cb827d96a3b1cb4c859adbaefd0ff246ac8c72d235de7"} Nov 25 05:51:10 crc kubenswrapper[4708]: I1125 05:51:10.156220 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:10 crc kubenswrapper[4708]: I1125 05:51:10.158808 4708 generic.go:334] "Generic (PLEG): container finished" podID="415c292a-1706-4c7b-9f9e-30f469708872" containerID="badeb684d2db1246b3e625cd1acac97a658acb9b06b3992383bba2e2979ced1b" exitCode=0 Nov 25 05:51:10 crc kubenswrapper[4708]: I1125 05:51:10.158897 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerDied","Data":"badeb684d2db1246b3e625cd1acac97a658acb9b06b3992383bba2e2979ced1b"} Nov 25 05:51:10 crc kubenswrapper[4708]: I1125 05:51:10.172319 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" podStartSLOduration=1.9751894060000001 podStartE2EDuration="8.172305972s" podCreationTimestamp="2025-11-25 05:51:02 +0000 UTC" firstStartedPulling="2025-11-25 05:51:03.502167473 +0000 UTC m=+604.911000860" lastFinishedPulling="2025-11-25 05:51:09.699284041 +0000 UTC m=+611.108117426" observedRunningTime="2025-11-25 05:51:10.169666979 +0000 UTC m=+611.578500366" watchObservedRunningTime="2025-11-25 05:51:10.172305972 +0000 UTC m=+611.581139357" Nov 25 05:51:11 crc kubenswrapper[4708]: I1125 05:51:11.169701 4708 generic.go:334] "Generic (PLEG): container finished" podID="415c292a-1706-4c7b-9f9e-30f469708872" containerID="074cf7b711358a13b6b09035b19473b5223c09d56421c6fba8dffffb531fef8d" exitCode=0 Nov 25 05:51:11 crc kubenswrapper[4708]: I1125 05:51:11.169808 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerDied","Data":"074cf7b711358a13b6b09035b19473b5223c09d56421c6fba8dffffb531fef8d"} Nov 25 05:51:12 crc kubenswrapper[4708]: I1125 05:51:12.180931 4708 generic.go:334] "Generic (PLEG): container finished" podID="415c292a-1706-4c7b-9f9e-30f469708872" containerID="786313ba0828df0d8997780ec49f9ec48806e5e2efd1d618173b5729932d5dea" exitCode=0 Nov 25 05:51:12 crc kubenswrapper[4708]: I1125 05:51:12.181017 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerDied","Data":"786313ba0828df0d8997780ec49f9ec48806e5e2efd1d618173b5729932d5dea"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.173761 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-bkbdq" Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194330 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"3546663e27072a503818a29615381ef34f31bd8a9a9809c894cfa3769a816622"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194364 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"fedc2ae11b728cbfb65bc442797866a44defe6eda44c22593c224a7241ad19df"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194375 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"7801e47942c68648fb59c868f7ea293cba27a6334090059e74537fef9670e498"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194385 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"6e1d55e2a00f03c74a67ae37e952fc3c045dec7c51bf07f1d8690ee7afd982b5"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"a2a8e4739b103946e5a262708d481abe13f0c5ff9a32b38658790f735189d8d6"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194400 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-f8lcr" event={"ID":"415c292a-1706-4c7b-9f9e-30f469708872","Type":"ContainerStarted","Data":"946eeb524dce3757c62f8f81bc0b970a6670d22fec091fb49f528f9ed6bcd875"} Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.194512 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.225346 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-f8lcr" podStartSLOduration=4.749761086 podStartE2EDuration="11.225327573s" podCreationTimestamp="2025-11-25 05:51:02 +0000 UTC" firstStartedPulling="2025-11-25 05:51:03.206442023 +0000 UTC m=+604.615275409" lastFinishedPulling="2025-11-25 05:51:09.682008509 +0000 UTC m=+611.090841896" observedRunningTime="2025-11-25 05:51:13.221862254 +0000 UTC m=+614.630695640" watchObservedRunningTime="2025-11-25 05:51:13.225327573 +0000 UTC m=+614.634160960" Nov 25 05:51:13 crc kubenswrapper[4708]: I1125 05:51:13.759084 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-dg7zk" Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.851944 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.852819 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.854487 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-27bkp" Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.860804 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.860938 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 05:51:15 crc kubenswrapper[4708]: I1125 05:51:15.865287 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:16 crc kubenswrapper[4708]: I1125 05:51:16.025306 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5n2c\" (UniqueName: \"kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c\") pod \"openstack-operator-index-mqvgc\" (UID: \"3c80a453-6319-4848-a96f-625dcc28f7f1\") " pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:16 crc kubenswrapper[4708]: I1125 05:51:16.126683 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5n2c\" (UniqueName: \"kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c\") pod \"openstack-operator-index-mqvgc\" (UID: \"3c80a453-6319-4848-a96f-625dcc28f7f1\") " pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:16 crc kubenswrapper[4708]: I1125 05:51:16.141493 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5n2c\" (UniqueName: \"kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c\") pod \"openstack-operator-index-mqvgc\" (UID: \"3c80a453-6319-4848-a96f-625dcc28f7f1\") " pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:16 crc kubenswrapper[4708]: I1125 05:51:16.167284 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:16 crc kubenswrapper[4708]: I1125 05:51:16.510537 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:16 crc kubenswrapper[4708]: W1125 05:51:16.515649 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c80a453_6319_4848_a96f_625dcc28f7f1.slice/crio-f3358aae81c7f74b69ddd3d5ef7aa1474ff6deef87125820b1569c4506ecf730 WatchSource:0}: Error finding container f3358aae81c7f74b69ddd3d5ef7aa1474ff6deef87125820b1569c4506ecf730: Status 404 returned error can't find the container with id f3358aae81c7f74b69ddd3d5ef7aa1474ff6deef87125820b1569c4506ecf730 Nov 25 05:51:17 crc kubenswrapper[4708]: I1125 05:51:17.215376 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqvgc" event={"ID":"3c80a453-6319-4848-a96f-625dcc28f7f1","Type":"ContainerStarted","Data":"f3358aae81c7f74b69ddd3d5ef7aa1474ff6deef87125820b1569c4506ecf730"} Nov 25 05:51:18 crc kubenswrapper[4708]: I1125 05:51:18.088689 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:18 crc kubenswrapper[4708]: I1125 05:51:18.135830 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:18 crc kubenswrapper[4708]: I1125 05:51:18.222166 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqvgc" event={"ID":"3c80a453-6319-4848-a96f-625dcc28f7f1","Type":"ContainerStarted","Data":"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe"} Nov 25 05:51:18 crc kubenswrapper[4708]: I1125 05:51:18.236723 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-mqvgc" podStartSLOduration=2.006439597 podStartE2EDuration="3.236710724s" podCreationTimestamp="2025-11-25 05:51:15 +0000 UTC" firstStartedPulling="2025-11-25 05:51:16.517591467 +0000 UTC m=+617.926424853" lastFinishedPulling="2025-11-25 05:51:17.747862593 +0000 UTC m=+619.156695980" observedRunningTime="2025-11-25 05:51:18.233109468 +0000 UTC m=+619.641942855" watchObservedRunningTime="2025-11-25 05:51:18.236710724 +0000 UTC m=+619.645544111" Nov 25 05:51:19 crc kubenswrapper[4708]: I1125 05:51:19.231993 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:19 crc kubenswrapper[4708]: I1125 05:51:19.836768 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-zkbgb"] Nov 25 05:51:19 crc kubenswrapper[4708]: I1125 05:51:19.837796 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:19 crc kubenswrapper[4708]: I1125 05:51:19.844678 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zkbgb"] Nov 25 05:51:19 crc kubenswrapper[4708]: I1125 05:51:19.975645 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnt8l\" (UniqueName: \"kubernetes.io/projected/7dc7a5c4-2123-4ca4-ad42-364233f79bb5-kube-api-access-xnt8l\") pod \"openstack-operator-index-zkbgb\" (UID: \"7dc7a5c4-2123-4ca4-ad42-364233f79bb5\") " pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.077017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnt8l\" (UniqueName: \"kubernetes.io/projected/7dc7a5c4-2123-4ca4-ad42-364233f79bb5-kube-api-access-xnt8l\") pod \"openstack-operator-index-zkbgb\" (UID: \"7dc7a5c4-2123-4ca4-ad42-364233f79bb5\") " pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.092503 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnt8l\" (UniqueName: \"kubernetes.io/projected/7dc7a5c4-2123-4ca4-ad42-364233f79bb5-kube-api-access-xnt8l\") pod \"openstack-operator-index-zkbgb\" (UID: \"7dc7a5c4-2123-4ca4-ad42-364233f79bb5\") " pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.151382 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.235254 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-mqvgc" podUID="3c80a453-6319-4848-a96f-625dcc28f7f1" containerName="registry-server" containerID="cri-o://95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe" gracePeriod=2 Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.500691 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-zkbgb"] Nov 25 05:51:20 crc kubenswrapper[4708]: W1125 05:51:20.505300 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dc7a5c4_2123_4ca4_ad42_364233f79bb5.slice/crio-bb2c1520960ff24f3fecdd47838137daae3d6f1535658f170db4fd9ef94aded4 WatchSource:0}: Error finding container bb2c1520960ff24f3fecdd47838137daae3d6f1535658f170db4fd9ef94aded4: Status 404 returned error can't find the container with id bb2c1520960ff24f3fecdd47838137daae3d6f1535658f170db4fd9ef94aded4 Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.525230 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.684652 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5n2c\" (UniqueName: \"kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c\") pod \"3c80a453-6319-4848-a96f-625dcc28f7f1\" (UID: \"3c80a453-6319-4848-a96f-625dcc28f7f1\") " Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.689311 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c" (OuterVolumeSpecName: "kube-api-access-s5n2c") pod "3c80a453-6319-4848-a96f-625dcc28f7f1" (UID: "3c80a453-6319-4848-a96f-625dcc28f7f1"). InnerVolumeSpecName "kube-api-access-s5n2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:51:20 crc kubenswrapper[4708]: I1125 05:51:20.786319 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5n2c\" (UniqueName: \"kubernetes.io/projected/3c80a453-6319-4848-a96f-625dcc28f7f1-kube-api-access-s5n2c\") on node \"crc\" DevicePath \"\"" Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.241843 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zkbgb" event={"ID":"7dc7a5c4-2123-4ca4-ad42-364233f79bb5","Type":"ContainerStarted","Data":"42ea141ad6e651dcb1bcbf8e5576018abb2229c328719822b6ac2138c43081b8"} Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.241894 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-zkbgb" event={"ID":"7dc7a5c4-2123-4ca4-ad42-364233f79bb5","Type":"ContainerStarted","Data":"bb2c1520960ff24f3fecdd47838137daae3d6f1535658f170db4fd9ef94aded4"} Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.243367 4708 generic.go:334] "Generic (PLEG): container finished" podID="3c80a453-6319-4848-a96f-625dcc28f7f1" containerID="95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe" exitCode=0 Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.243440 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqvgc" event={"ID":"3c80a453-6319-4848-a96f-625dcc28f7f1","Type":"ContainerDied","Data":"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe"} Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.243467 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-mqvgc" event={"ID":"3c80a453-6319-4848-a96f-625dcc28f7f1","Type":"ContainerDied","Data":"f3358aae81c7f74b69ddd3d5ef7aa1474ff6deef87125820b1569c4506ecf730"} Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.243486 4708 scope.go:117] "RemoveContainer" containerID="95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe" Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.243493 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-mqvgc" Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.256678 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-zkbgb" podStartSLOduration=1.746385895 podStartE2EDuration="2.256661912s" podCreationTimestamp="2025-11-25 05:51:19 +0000 UTC" firstStartedPulling="2025-11-25 05:51:20.509799098 +0000 UTC m=+621.918632483" lastFinishedPulling="2025-11-25 05:51:21.020075114 +0000 UTC m=+622.428908500" observedRunningTime="2025-11-25 05:51:21.252210414 +0000 UTC m=+622.661043800" watchObservedRunningTime="2025-11-25 05:51:21.256661912 +0000 UTC m=+622.665495298" Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.264365 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.267075 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-mqvgc"] Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.771979 4708 scope.go:117] "RemoveContainer" containerID="95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe" Nov 25 05:51:21 crc kubenswrapper[4708]: E1125 05:51:21.772452 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe\": container with ID starting with 95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe not found: ID does not exist" containerID="95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe" Nov 25 05:51:21 crc kubenswrapper[4708]: I1125 05:51:21.772501 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe"} err="failed to get container status \"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe\": rpc error: code = NotFound desc = could not find container \"95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe\": container with ID starting with 95a98e10d23f56217468e729dac059c2fc0b27465e27b02787dcd027557482fe not found: ID does not exist" Nov 25 05:51:22 crc kubenswrapper[4708]: I1125 05:51:22.899252 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c80a453-6319-4848-a96f-625dcc28f7f1" path="/var/lib/kubelet/pods/3c80a453-6319-4848-a96f-625dcc28f7f1/volumes" Nov 25 05:51:23 crc kubenswrapper[4708]: I1125 05:51:23.092662 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-f8lcr" Nov 25 05:51:23 crc kubenswrapper[4708]: I1125 05:51:23.110045 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-5fbpc" Nov 25 05:51:30 crc kubenswrapper[4708]: I1125 05:51:30.151762 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:30 crc kubenswrapper[4708]: I1125 05:51:30.152315 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:30 crc kubenswrapper[4708]: I1125 05:51:30.174981 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:30 crc kubenswrapper[4708]: I1125 05:51:30.325698 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-zkbgb" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.174574 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g"] Nov 25 05:51:35 crc kubenswrapper[4708]: E1125 05:51:35.175359 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c80a453-6319-4848-a96f-625dcc28f7f1" containerName="registry-server" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.175373 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c80a453-6319-4848-a96f-625dcc28f7f1" containerName="registry-server" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.175483 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c80a453-6319-4848-a96f-625dcc28f7f1" containerName="registry-server" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.176311 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.178973 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-bvpjg" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.182814 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g"] Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.274341 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.274393 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.274430 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9jpm\" (UniqueName: \"kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.375672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.375952 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.376009 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9jpm\" (UniqueName: \"kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.376195 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.376484 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.391758 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9jpm\" (UniqueName: \"kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm\") pod \"bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.491161 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:35 crc kubenswrapper[4708]: I1125 05:51:35.863033 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g"] Nov 25 05:51:35 crc kubenswrapper[4708]: W1125 05:51:35.867376 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod322a79ef_4a89_4999_a433_93fc72601580.slice/crio-a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e WatchSource:0}: Error finding container a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e: Status 404 returned error can't find the container with id a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e Nov 25 05:51:36 crc kubenswrapper[4708]: I1125 05:51:36.333335 4708 generic.go:334] "Generic (PLEG): container finished" podID="322a79ef-4a89-4999-a433-93fc72601580" containerID="b965d4690e474c519b7fdc3609a345fb13390de745dbc4b3ba1d36fd91d1afa8" exitCode=0 Nov 25 05:51:36 crc kubenswrapper[4708]: I1125 05:51:36.333371 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" event={"ID":"322a79ef-4a89-4999-a433-93fc72601580","Type":"ContainerDied","Data":"b965d4690e474c519b7fdc3609a345fb13390de745dbc4b3ba1d36fd91d1afa8"} Nov 25 05:51:36 crc kubenswrapper[4708]: I1125 05:51:36.333627 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" event={"ID":"322a79ef-4a89-4999-a433-93fc72601580","Type":"ContainerStarted","Data":"a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e"} Nov 25 05:51:38 crc kubenswrapper[4708]: I1125 05:51:38.351289 4708 generic.go:334] "Generic (PLEG): container finished" podID="322a79ef-4a89-4999-a433-93fc72601580" containerID="c0e7c3f98f0ab9963459101e74428c7f0697101707b5f9eba8c41d007dec09d3" exitCode=0 Nov 25 05:51:38 crc kubenswrapper[4708]: I1125 05:51:38.351373 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" event={"ID":"322a79ef-4a89-4999-a433-93fc72601580","Type":"ContainerDied","Data":"c0e7c3f98f0ab9963459101e74428c7f0697101707b5f9eba8c41d007dec09d3"} Nov 25 05:51:39 crc kubenswrapper[4708]: I1125 05:51:39.364035 4708 generic.go:334] "Generic (PLEG): container finished" podID="322a79ef-4a89-4999-a433-93fc72601580" containerID="2c9efacf8317ac4ff3d5e94ff0d4dbd3f2dbbd6c065900657838c63cbe446cc6" exitCode=0 Nov 25 05:51:39 crc kubenswrapper[4708]: I1125 05:51:39.364120 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" event={"ID":"322a79ef-4a89-4999-a433-93fc72601580","Type":"ContainerDied","Data":"2c9efacf8317ac4ff3d5e94ff0d4dbd3f2dbbd6c065900657838c63cbe446cc6"} Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.568707 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.752408 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util\") pod \"322a79ef-4a89-4999-a433-93fc72601580\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.752467 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9jpm\" (UniqueName: \"kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm\") pod \"322a79ef-4a89-4999-a433-93fc72601580\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.753485 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle\") pod \"322a79ef-4a89-4999-a433-93fc72601580\" (UID: \"322a79ef-4a89-4999-a433-93fc72601580\") " Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.754190 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle" (OuterVolumeSpecName: "bundle") pod "322a79ef-4a89-4999-a433-93fc72601580" (UID: "322a79ef-4a89-4999-a433-93fc72601580"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.759406 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm" (OuterVolumeSpecName: "kube-api-access-k9jpm") pod "322a79ef-4a89-4999-a433-93fc72601580" (UID: "322a79ef-4a89-4999-a433-93fc72601580"). InnerVolumeSpecName "kube-api-access-k9jpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.763074 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util" (OuterVolumeSpecName: "util") pod "322a79ef-4a89-4999-a433-93fc72601580" (UID: "322a79ef-4a89-4999-a433-93fc72601580"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.854804 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9jpm\" (UniqueName: \"kubernetes.io/projected/322a79ef-4a89-4999-a433-93fc72601580-kube-api-access-k9jpm\") on node \"crc\" DevicePath \"\"" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.855010 4708 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-util\") on node \"crc\" DevicePath \"\"" Nov 25 05:51:40 crc kubenswrapper[4708]: I1125 05:51:40.855081 4708 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/322a79ef-4a89-4999-a433-93fc72601580-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:51:41 crc kubenswrapper[4708]: I1125 05:51:41.381162 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" event={"ID":"322a79ef-4a89-4999-a433-93fc72601580","Type":"ContainerDied","Data":"a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e"} Nov 25 05:51:41 crc kubenswrapper[4708]: I1125 05:51:41.381224 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a443d48e0d236ec7b7e7f60f378178e281d8c402d7f68e95168080738448914e" Nov 25 05:51:41 crc kubenswrapper[4708]: I1125 05:51:41.381311 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.351251 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh"] Nov 25 05:51:47 crc kubenswrapper[4708]: E1125 05:51:47.351676 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="util" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.351689 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="util" Nov 25 05:51:47 crc kubenswrapper[4708]: E1125 05:51:47.351702 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="extract" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.351709 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="extract" Nov 25 05:51:47 crc kubenswrapper[4708]: E1125 05:51:47.351724 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="pull" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.351729 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="pull" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.351818 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="322a79ef-4a89-4999-a433-93fc72601580" containerName="extract" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.352159 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.353731 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-mpmqt" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.376968 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh"] Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.441179 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k4hj\" (UniqueName: \"kubernetes.io/projected/5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1-kube-api-access-2k4hj\") pod \"openstack-operator-controller-operator-7b567956b5-lb9hh\" (UID: \"5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.542596 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k4hj\" (UniqueName: \"kubernetes.io/projected/5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1-kube-api-access-2k4hj\") pod \"openstack-operator-controller-operator-7b567956b5-lb9hh\" (UID: \"5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.559786 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k4hj\" (UniqueName: \"kubernetes.io/projected/5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1-kube-api-access-2k4hj\") pod \"openstack-operator-controller-operator-7b567956b5-lb9hh\" (UID: \"5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1\") " pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:47 crc kubenswrapper[4708]: I1125 05:51:47.666676 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:48 crc kubenswrapper[4708]: I1125 05:51:48.089379 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh"] Nov 25 05:51:48 crc kubenswrapper[4708]: I1125 05:51:48.427795 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" event={"ID":"5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1","Type":"ContainerStarted","Data":"07155eb5d6c39f7f64882c0d6793d9ab73d852b9ad373ecf755d7113b2695f77"} Nov 25 05:51:53 crc kubenswrapper[4708]: I1125 05:51:53.475096 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:51:53 crc kubenswrapper[4708]: I1125 05:51:53.501952 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" podStartSLOduration=1.254018978 podStartE2EDuration="6.50192914s" podCreationTimestamp="2025-11-25 05:51:47 +0000 UTC" firstStartedPulling="2025-11-25 05:51:48.097757905 +0000 UTC m=+649.506591290" lastFinishedPulling="2025-11-25 05:51:53.345668066 +0000 UTC m=+654.754501452" observedRunningTime="2025-11-25 05:51:53.49822947 +0000 UTC m=+654.907062856" watchObservedRunningTime="2025-11-25 05:51:53.50192914 +0000 UTC m=+654.910762527" Nov 25 05:51:54 crc kubenswrapper[4708]: I1125 05:51:54.483490 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" event={"ID":"5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1","Type":"ContainerStarted","Data":"8491e85cc6a40fde775f821f0b360858b251043b3af28d4c3791140bda17fcf9"} Nov 25 05:52:07 crc kubenswrapper[4708]: I1125 05:52:07.669909 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7b567956b5-lb9hh" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.509823 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.511229 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.514187 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-zp47g" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.518490 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.520459 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.541665 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-gpqff" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.554882 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.566572 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.567601 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.568125 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.577982 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-8f2b7" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.578264 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w9xx\" (UniqueName: \"kubernetes.io/projected/8cb5eab2-d137-4aeb-9ca5-59139c67a2de-kube-api-access-7w9xx\") pod \"barbican-operator-controller-manager-86dc4d89c8-kvrrg\" (UID: \"8cb5eab2-d137-4aeb-9ca5-59139c67a2de\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.578406 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lch7\" (UniqueName: \"kubernetes.io/projected/13a9d474-f528-4913-ada4-313db91a37d5-kube-api-access-2lch7\") pod \"cinder-operator-controller-manager-79856dc55c-b7hd2\" (UID: \"13a9d474-f528-4913-ada4-313db91a37d5\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.578514 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5v9tj\" (UniqueName: \"kubernetes.io/projected/fd2a735a-3219-45d8-a689-1a8722923d8d-kube-api-access-5v9tj\") pod \"designate-operator-controller-manager-7d695c9b56-j2vk8\" (UID: \"fd2a735a-3219-45d8-a689-1a8722923d8d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.600610 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.601900 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.604661 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.609406 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4qhkm" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.617087 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-9knwb"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.618120 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.623160 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-zh2pb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.628937 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.640048 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-9knwb"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.646490 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.647595 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.660403 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-22wgs" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.662837 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.663946 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.666508 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jd9ts" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.666746 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679467 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lch7\" (UniqueName: \"kubernetes.io/projected/13a9d474-f528-4913-ada4-313db91a37d5-kube-api-access-2lch7\") pod \"cinder-operator-controller-manager-79856dc55c-b7hd2\" (UID: \"13a9d474-f528-4913-ada4-313db91a37d5\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679556 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w4z4\" (UniqueName: \"kubernetes.io/projected/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-kube-api-access-6w4z4\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679597 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679649 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5v9tj\" (UniqueName: \"kubernetes.io/projected/fd2a735a-3219-45d8-a689-1a8722923d8d-kube-api-access-5v9tj\") pod \"designate-operator-controller-manager-7d695c9b56-j2vk8\" (UID: \"fd2a735a-3219-45d8-a689-1a8722923d8d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679695 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm5df\" (UniqueName: \"kubernetes.io/projected/b752b0d3-bffb-4322-9e11-d2a9451c4f39-kube-api-access-qm5df\") pod \"horizon-operator-controller-manager-68c9694994-j98rt\" (UID: \"b752b0d3-bffb-4322-9e11-d2a9451c4f39\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679718 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w9xx\" (UniqueName: \"kubernetes.io/projected/8cb5eab2-d137-4aeb-9ca5-59139c67a2de-kube-api-access-7w9xx\") pod \"barbican-operator-controller-manager-86dc4d89c8-kvrrg\" (UID: \"8cb5eab2-d137-4aeb-9ca5-59139c67a2de\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679738 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htksq\" (UniqueName: \"kubernetes.io/projected/e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2-kube-api-access-htksq\") pod \"heat-operator-controller-manager-774b86978c-9knwb\" (UID: \"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.679767 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jh2f\" (UniqueName: \"kubernetes.io/projected/eef4d5f6-a70a-4b08-8836-3dde6ab0da36-kube-api-access-7jh2f\") pod \"glance-operator-controller-manager-68b95954c9-df8zr\" (UID: \"eef4d5f6-a70a-4b08-8836-3dde6ab0da36\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.690608 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.699288 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.702374 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.703430 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.710362 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w9xx\" (UniqueName: \"kubernetes.io/projected/8cb5eab2-d137-4aeb-9ca5-59139c67a2de-kube-api-access-7w9xx\") pod \"barbican-operator-controller-manager-86dc4d89c8-kvrrg\" (UID: \"8cb5eab2-d137-4aeb-9ca5-59139c67a2de\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.715090 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lch7\" (UniqueName: \"kubernetes.io/projected/13a9d474-f528-4913-ada4-313db91a37d5-kube-api-access-2lch7\") pod \"cinder-operator-controller-manager-79856dc55c-b7hd2\" (UID: \"13a9d474-f528-4913-ada4-313db91a37d5\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.716589 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2cgnq" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.717853 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.718775 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.721756 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5v9tj\" (UniqueName: \"kubernetes.io/projected/fd2a735a-3219-45d8-a689-1a8722923d8d-kube-api-access-5v9tj\") pod \"designate-operator-controller-manager-7d695c9b56-j2vk8\" (UID: \"fd2a735a-3219-45d8-a689-1a8722923d8d\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.725672 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-8crmf" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.726564 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.733140 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.742546 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.743677 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.751220 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qjx9f" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.751788 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.752708 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.762218 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-82whq" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.770716 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.771833 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.778312 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-rmbdd" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.778490 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.783453 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784228 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm5df\" (UniqueName: \"kubernetes.io/projected/b752b0d3-bffb-4322-9e11-d2a9451c4f39-kube-api-access-qm5df\") pod \"horizon-operator-controller-manager-68c9694994-j98rt\" (UID: \"b752b0d3-bffb-4322-9e11-d2a9451c4f39\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784262 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5tzf\" (UniqueName: \"kubernetes.io/projected/b2240efc-4688-4d79-a19d-ea3801742494-kube-api-access-x5tzf\") pod \"manila-operator-controller-manager-58bb8d67cc-6zc24\" (UID: \"b2240efc-4688-4d79-a19d-ea3801742494\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784290 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htksq\" (UniqueName: \"kubernetes.io/projected/e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2-kube-api-access-htksq\") pod \"heat-operator-controller-manager-774b86978c-9knwb\" (UID: \"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784320 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jh2f\" (UniqueName: \"kubernetes.io/projected/eef4d5f6-a70a-4b08-8836-3dde6ab0da36-kube-api-access-7jh2f\") pod \"glance-operator-controller-manager-68b95954c9-df8zr\" (UID: \"eef4d5f6-a70a-4b08-8836-3dde6ab0da36\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784348 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x7wv\" (UniqueName: \"kubernetes.io/projected/1e34dcae-a471-4c9c-83f2-43d551516520-kube-api-access-9x7wv\") pod \"keystone-operator-controller-manager-748dc6576f-9wnpj\" (UID: \"1e34dcae-a471-4c9c-83f2-43d551516520\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784375 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwzqt\" (UniqueName: \"kubernetes.io/projected/17ef75be-981e-4579-ae1b-aa7e4c4b8918-kube-api-access-xwzqt\") pod \"neutron-operator-controller-manager-7c57c8bbc4-c8c5l\" (UID: \"17ef75be-981e-4579-ae1b-aa7e4c4b8918\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784395 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sqk2\" (UniqueName: \"kubernetes.io/projected/1fd35d81-f228-4c45-895a-95feb523ef1f-kube-api-access-2sqk2\") pod \"ironic-operator-controller-manager-5bfcdc958c-8xsh6\" (UID: \"1fd35d81-f228-4c45-895a-95feb523ef1f\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784447 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w4z4\" (UniqueName: \"kubernetes.io/projected/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-kube-api-access-6w4z4\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784476 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.784500 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf8zd\" (UniqueName: \"kubernetes.io/projected/856fc8ab-76a7-4ed4-ab1e-31dad3985982-kube-api-access-jf8zd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-x4xzv\" (UID: \"856fc8ab-76a7-4ed4-ab1e-31dad3985982\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:24 crc kubenswrapper[4708]: E1125 05:52:24.785158 4708 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 05:52:24 crc kubenswrapper[4708]: E1125 05:52:24.785207 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert podName:49cb03d2-7430-4f92-b82b-7e4b7ad9faee nodeName:}" failed. No retries permitted until 2025-11-25 05:52:25.285191967 +0000 UTC m=+686.694025353 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert") pod "infra-operator-controller-manager-d5cc86f4b-j5g96" (UID: "49cb03d2-7430-4f92-b82b-7e4b7ad9faee") : secret "infra-operator-webhook-server-cert" not found Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.785841 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.786844 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.793137 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-m8mff" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.798570 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.802574 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htksq\" (UniqueName: \"kubernetes.io/projected/e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2-kube-api-access-htksq\") pod \"heat-operator-controller-manager-774b86978c-9knwb\" (UID: \"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.806965 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w4z4\" (UniqueName: \"kubernetes.io/projected/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-kube-api-access-6w4z4\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.807040 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm5df\" (UniqueName: \"kubernetes.io/projected/b752b0d3-bffb-4322-9e11-d2a9451c4f39-kube-api-access-qm5df\") pod \"horizon-operator-controller-manager-68c9694994-j98rt\" (UID: \"b752b0d3-bffb-4322-9e11-d2a9451c4f39\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.811178 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jh2f\" (UniqueName: \"kubernetes.io/projected/eef4d5f6-a70a-4b08-8836-3dde6ab0da36-kube-api-access-7jh2f\") pod \"glance-operator-controller-manager-68b95954c9-df8zr\" (UID: \"eef4d5f6-a70a-4b08-8836-3dde6ab0da36\") " pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.818340 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.843879 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.862956 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.880658 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.888764 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891012 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x7wv\" (UniqueName: \"kubernetes.io/projected/1e34dcae-a471-4c9c-83f2-43d551516520-kube-api-access-9x7wv\") pod \"keystone-operator-controller-manager-748dc6576f-9wnpj\" (UID: \"1e34dcae-a471-4c9c-83f2-43d551516520\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891077 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwzqt\" (UniqueName: \"kubernetes.io/projected/17ef75be-981e-4579-ae1b-aa7e4c4b8918-kube-api-access-xwzqt\") pod \"neutron-operator-controller-manager-7c57c8bbc4-c8c5l\" (UID: \"17ef75be-981e-4579-ae1b-aa7e4c4b8918\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891108 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sqk2\" (UniqueName: \"kubernetes.io/projected/1fd35d81-f228-4c45-895a-95feb523ef1f-kube-api-access-2sqk2\") pod \"ironic-operator-controller-manager-5bfcdc958c-8xsh6\" (UID: \"1fd35d81-f228-4c45-895a-95feb523ef1f\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891195 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf8zd\" (UniqueName: \"kubernetes.io/projected/856fc8ab-76a7-4ed4-ab1e-31dad3985982-kube-api-access-jf8zd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-x4xzv\" (UID: \"856fc8ab-76a7-4ed4-ab1e-31dad3985982\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891265 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5tzf\" (UniqueName: \"kubernetes.io/projected/b2240efc-4688-4d79-a19d-ea3801742494-kube-api-access-x5tzf\") pod \"manila-operator-controller-manager-58bb8d67cc-6zc24\" (UID: \"b2240efc-4688-4d79-a19d-ea3801742494\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.891306 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg26z\" (UniqueName: \"kubernetes.io/projected/dcbf8314-b18c-43ce-8a2e-400b9da62852-kube-api-access-kg26z\") pod \"nova-operator-controller-manager-79556f57fc-qjcbh\" (UID: \"dcbf8314-b18c-43ce-8a2e-400b9da62852\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.917539 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.922725 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf8zd\" (UniqueName: \"kubernetes.io/projected/856fc8ab-76a7-4ed4-ab1e-31dad3985982-kube-api-access-jf8zd\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-x4xzv\" (UID: \"856fc8ab-76a7-4ed4-ab1e-31dad3985982\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.923862 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-qljwr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.924781 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.930059 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x7wv\" (UniqueName: \"kubernetes.io/projected/1e34dcae-a471-4c9c-83f2-43d551516520-kube-api-access-9x7wv\") pod \"keystone-operator-controller-manager-748dc6576f-9wnpj\" (UID: \"1e34dcae-a471-4c9c-83f2-43d551516520\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.933056 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sqk2\" (UniqueName: \"kubernetes.io/projected/1fd35d81-f228-4c45-895a-95feb523ef1f-kube-api-access-2sqk2\") pod \"ironic-operator-controller-manager-5bfcdc958c-8xsh6\" (UID: \"1fd35d81-f228-4c45-895a-95feb523ef1f\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.933727 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5tzf\" (UniqueName: \"kubernetes.io/projected/b2240efc-4688-4d79-a19d-ea3801742494-kube-api-access-x5tzf\") pod \"manila-operator-controller-manager-58bb8d67cc-6zc24\" (UID: \"b2240efc-4688-4d79-a19d-ea3801742494\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.942771 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.951323 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwzqt\" (UniqueName: \"kubernetes.io/projected/17ef75be-981e-4579-ae1b-aa7e4c4b8918-kube-api-access-xwzqt\") pod \"neutron-operator-controller-manager-7c57c8bbc4-c8c5l\" (UID: \"17ef75be-981e-4579-ae1b-aa7e4c4b8918\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.975859 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.977005 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.977786 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.977809 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.977827 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.977905 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.978088 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.979027 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.979450 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.979495 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.979510 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.980332 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.980468 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.980550 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.982696 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-k6827" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.982873 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.982972 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-6dkvb" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.983089 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-8gpk5" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.985126 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.986036 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.987032 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-nblzm" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.990559 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66"] Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.993652 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wslns\" (UniqueName: \"kubernetes.io/projected/a7c6b74d-a672-4349-b9f7-52a726bc8ec9-kube-api-access-wslns\") pod \"ovn-operator-controller-manager-66cf5c67ff-gnn9s\" (UID: \"a7c6b74d-a672-4349-b9f7-52a726bc8ec9\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.993692 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxzrz\" (UniqueName: \"kubernetes.io/projected/3d11487d-d74d-428d-b18b-800238416691-kube-api-access-cxzrz\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.993727 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg26z\" (UniqueName: \"kubernetes.io/projected/dcbf8314-b18c-43ce-8a2e-400b9da62852-kube-api-access-kg26z\") pod \"nova-operator-controller-manager-79556f57fc-qjcbh\" (UID: \"dcbf8314-b18c-43ce-8a2e-400b9da62852\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.993796 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.994036 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2t2l\" (UniqueName: \"kubernetes.io/projected/6af5f14d-1c9e-4985-ba46-3f4b193ac96d-kube-api-access-l2t2l\") pod \"octavia-operator-controller-manager-fd75fd47d-rmmvr\" (UID: \"6af5f14d-1c9e-4985-ba46-3f4b193ac96d\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:24 crc kubenswrapper[4708]: I1125 05:52:24.995031 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-9hrt8" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.035411 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg26z\" (UniqueName: \"kubernetes.io/projected/dcbf8314-b18c-43ce-8a2e-400b9da62852-kube-api-access-kg26z\") pod \"nova-operator-controller-manager-79556f57fc-qjcbh\" (UID: \"dcbf8314-b18c-43ce-8a2e-400b9da62852\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.057821 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.070676 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.072487 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.073552 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.080692 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.083339 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.085293 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-7gd48" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.094871 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.095953 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkb8c\" (UniqueName: \"kubernetes.io/projected/83077424-8cf5-48ae-b8b2-d0b94fb650e6-kube-api-access-xkb8c\") pod \"swift-operator-controller-manager-6fdc4fcf86-d64tb\" (UID: \"83077424-8cf5-48ae-b8b2-d0b94fb650e6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096028 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2t2l\" (UniqueName: \"kubernetes.io/projected/6af5f14d-1c9e-4985-ba46-3f4b193ac96d-kube-api-access-l2t2l\") pod \"octavia-operator-controller-manager-fd75fd47d-rmmvr\" (UID: \"6af5f14d-1c9e-4985-ba46-3f4b193ac96d\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096051 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgfsn\" (UniqueName: \"kubernetes.io/projected/1064823e-17fb-42bc-b68e-cfab481fbf8c-kube-api-access-mgfsn\") pod \"placement-operator-controller-manager-5db546f9d9-vzz64\" (UID: \"1064823e-17fb-42bc-b68e-cfab481fbf8c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096095 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wslns\" (UniqueName: \"kubernetes.io/projected/a7c6b74d-a672-4349-b9f7-52a726bc8ec9-kube-api-access-wslns\") pod \"ovn-operator-controller-manager-66cf5c67ff-gnn9s\" (UID: \"a7c6b74d-a672-4349-b9f7-52a726bc8ec9\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096118 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxzrz\" (UniqueName: \"kubernetes.io/projected/3d11487d-d74d-428d-b18b-800238416691-kube-api-access-cxzrz\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096153 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8mjq\" (UniqueName: \"kubernetes.io/projected/7d60669a-b373-49bf-b65f-88b3e0d48d87-kube-api-access-g8mjq\") pod \"telemetry-operator-controller-manager-567f98c9d-phd66\" (UID: \"7d60669a-b373-49bf-b65f-88b3e0d48d87\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096181 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhb2j\" (UniqueName: \"kubernetes.io/projected/ed82966d-e1bb-448f-be61-091fcb5f3f2b-kube-api-access-jhb2j\") pod \"test-operator-controller-manager-5cb74df96-z9kvf\" (UID: \"ed82966d-e1bb-448f-be61-091fcb5f3f2b\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.096220 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.096338 4708 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.096394 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert podName:3d11487d-d74d-428d-b18b-800238416691 nodeName:}" failed. No retries permitted until 2025-11-25 05:52:25.596378718 +0000 UTC m=+687.005212103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" (UID: "3d11487d-d74d-428d-b18b-800238416691") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.105131 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.121937 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxzrz\" (UniqueName: \"kubernetes.io/projected/3d11487d-d74d-428d-b18b-800238416691-kube-api-access-cxzrz\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.122058 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2t2l\" (UniqueName: \"kubernetes.io/projected/6af5f14d-1c9e-4985-ba46-3f4b193ac96d-kube-api-access-l2t2l\") pod \"octavia-operator-controller-manager-fd75fd47d-rmmvr\" (UID: \"6af5f14d-1c9e-4985-ba46-3f4b193ac96d\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.122114 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wslns\" (UniqueName: \"kubernetes.io/projected/a7c6b74d-a672-4349-b9f7-52a726bc8ec9-kube-api-access-wslns\") pod \"ovn-operator-controller-manager-66cf5c67ff-gnn9s\" (UID: \"a7c6b74d-a672-4349-b9f7-52a726bc8ec9\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.137776 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-hhgb4"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.139347 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.141803 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-lqndc" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.147395 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.154806 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-hhgb4"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.197727 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkb8c\" (UniqueName: \"kubernetes.io/projected/83077424-8cf5-48ae-b8b2-d0b94fb650e6-kube-api-access-xkb8c\") pod \"swift-operator-controller-manager-6fdc4fcf86-d64tb\" (UID: \"83077424-8cf5-48ae-b8b2-d0b94fb650e6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.197803 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgfsn\" (UniqueName: \"kubernetes.io/projected/1064823e-17fb-42bc-b68e-cfab481fbf8c-kube-api-access-mgfsn\") pod \"placement-operator-controller-manager-5db546f9d9-vzz64\" (UID: \"1064823e-17fb-42bc-b68e-cfab481fbf8c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.197893 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8mjq\" (UniqueName: \"kubernetes.io/projected/7d60669a-b373-49bf-b65f-88b3e0d48d87-kube-api-access-g8mjq\") pod \"telemetry-operator-controller-manager-567f98c9d-phd66\" (UID: \"7d60669a-b373-49bf-b65f-88b3e0d48d87\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.197921 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz4fm\" (UniqueName: \"kubernetes.io/projected/cb75b241-90c4-45ee-9473-864b6b9f8705-kube-api-access-fz4fm\") pod \"watcher-operator-controller-manager-864885998-hhgb4\" (UID: \"cb75b241-90c4-45ee-9473-864b6b9f8705\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.197951 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhb2j\" (UniqueName: \"kubernetes.io/projected/ed82966d-e1bb-448f-be61-091fcb5f3f2b-kube-api-access-jhb2j\") pod \"test-operator-controller-manager-5cb74df96-z9kvf\" (UID: \"ed82966d-e1bb-448f-be61-091fcb5f3f2b\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.216469 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhb2j\" (UniqueName: \"kubernetes.io/projected/ed82966d-e1bb-448f-be61-091fcb5f3f2b-kube-api-access-jhb2j\") pod \"test-operator-controller-manager-5cb74df96-z9kvf\" (UID: \"ed82966d-e1bb-448f-be61-091fcb5f3f2b\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.227638 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8mjq\" (UniqueName: \"kubernetes.io/projected/7d60669a-b373-49bf-b65f-88b3e0d48d87-kube-api-access-g8mjq\") pod \"telemetry-operator-controller-manager-567f98c9d-phd66\" (UID: \"7d60669a-b373-49bf-b65f-88b3e0d48d87\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.228296 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkb8c\" (UniqueName: \"kubernetes.io/projected/83077424-8cf5-48ae-b8b2-d0b94fb650e6-kube-api-access-xkb8c\") pod \"swift-operator-controller-manager-6fdc4fcf86-d64tb\" (UID: \"83077424-8cf5-48ae-b8b2-d0b94fb650e6\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.231143 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgfsn\" (UniqueName: \"kubernetes.io/projected/1064823e-17fb-42bc-b68e-cfab481fbf8c-kube-api-access-mgfsn\") pod \"placement-operator-controller-manager-5db546f9d9-vzz64\" (UID: \"1064823e-17fb-42bc-b68e-cfab481fbf8c\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.257537 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.260053 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.267092 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.275209 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.276031 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-lkb67" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.276186 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.293610 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.299069 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6w8p\" (UniqueName: \"kubernetes.io/projected/af33b085-a912-42a9-88ba-6b7e8d27ec8d-kube-api-access-d6w8p\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.299114 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz4fm\" (UniqueName: \"kubernetes.io/projected/cb75b241-90c4-45ee-9473-864b6b9f8705-kube-api-access-fz4fm\") pod \"watcher-operator-controller-manager-864885998-hhgb4\" (UID: \"cb75b241-90c4-45ee-9473-864b6b9f8705\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.299139 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.299192 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.299229 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.302232 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/49cb03d2-7430-4f92-b82b-7e4b7ad9faee-cert\") pod \"infra-operator-controller-manager-d5cc86f4b-j5g96\" (UID: \"49cb03d2-7430-4f92-b82b-7e4b7ad9faee\") " pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.322129 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.342956 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz4fm\" (UniqueName: \"kubernetes.io/projected/cb75b241-90c4-45ee-9473-864b6b9f8705-kube-api-access-fz4fm\") pod \"watcher-operator-controller-manager-864885998-hhgb4\" (UID: \"cb75b241-90c4-45ee-9473-864b6b9f8705\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.345136 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.375117 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.375745 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.376832 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.379629 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-trlnj" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.387679 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.401900 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6w8p\" (UniqueName: \"kubernetes.io/projected/af33b085-a912-42a9-88ba-6b7e8d27ec8d-kube-api-access-d6w8p\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.401993 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.404688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.405190 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6qm7\" (UniqueName: \"kubernetes.io/projected/54a0ed25-ded8-471d-acdd-b34f75347701-kube-api-access-z6qm7\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mc72h\" (UID: \"54a0ed25-ded8-471d-acdd-b34f75347701\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.402955 4708 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.405389 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:25.905370102 +0000 UTC m=+687.314203489 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.405432 4708 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.405500 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:25.905481342 +0000 UTC m=+687.314314729 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "metrics-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.414777 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.421930 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6w8p\" (UniqueName: \"kubernetes.io/projected/af33b085-a912-42a9-88ba-6b7e8d27ec8d-kube-api-access-d6w8p\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.445313 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.464030 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.508401 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6qm7\" (UniqueName: \"kubernetes.io/projected/54a0ed25-ded8-471d-acdd-b34f75347701-kube-api-access-z6qm7\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mc72h\" (UID: \"54a0ed25-ded8-471d-acdd-b34f75347701\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.532365 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.536221 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.543386 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6qm7\" (UniqueName: \"kubernetes.io/projected/54a0ed25-ded8-471d-acdd-b34f75347701-kube-api-access-z6qm7\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mc72h\" (UID: \"54a0ed25-ded8-471d-acdd-b34f75347701\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.582947 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.610863 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.611067 4708 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.611120 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert podName:3d11487d-d74d-428d-b18b-800238416691 nodeName:}" failed. No retries permitted until 2025-11-25 05:52:26.611105205 +0000 UTC m=+688.019938580 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert") pod "openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" (UID: "3d11487d-d74d-428d-b18b-800238416691") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.643028 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8"] Nov 25 05:52:25 crc kubenswrapper[4708]: W1125 05:52:25.683069 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd2a735a_3219_45d8_a689_1a8722923d8d.slice/crio-ce96b7d811e16e020b9c9f8156071d35654c0ffaeed21ad3c458c8f76a9f42d1 WatchSource:0}: Error finding container ce96b7d811e16e020b9c9f8156071d35654c0ffaeed21ad3c458c8f76a9f42d1: Status 404 returned error can't find the container with id ce96b7d811e16e020b9c9f8156071d35654c0ffaeed21ad3c458c8f76a9f42d1 Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.717729 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.741401 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.743956 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" event={"ID":"fd2a735a-3219-45d8-a689-1a8722923d8d","Type":"ContainerStarted","Data":"ce96b7d811e16e020b9c9f8156071d35654c0ffaeed21ad3c458c8f76a9f42d1"} Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.768953 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" event={"ID":"13a9d474-f528-4913-ada4-313db91a37d5","Type":"ContainerStarted","Data":"53101784c2e82b6e3030f22bdff623058cc1ddfee1b591ec91a769c748c72385"} Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.773278 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" event={"ID":"8cb5eab2-d137-4aeb-9ca5-59139c67a2de","Type":"ContainerStarted","Data":"da7925ebe11306eaf5e304facfc18e986ec816b01bae4bd21e44c27f08874a46"} Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.773500 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.855917 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-9knwb"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.885236 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt"] Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.921003 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: I1125 05:52:25.921090 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.921206 4708 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.921297 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:26.921271873 +0000 UTC m=+688.330105260 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "webhook-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.921314 4708 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 05:52:25 crc kubenswrapper[4708]: E1125 05:52:25.921376 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:26.921353347 +0000 UTC m=+688.330186733 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "metrics-server-cert" not found Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.047710 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l"] Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.064762 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh"] Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.066669 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6"] Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.069960 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr"] Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.073039 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.086370 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6af5f14d_1c9e_4985_ba46_3f4b193ac96d.slice/crio-de24926ddfd3deb21b4ce351309ca6d639f9da4787bee9bb1a56765880606db0 WatchSource:0}: Error finding container de24926ddfd3deb21b4ce351309ca6d639f9da4787bee9bb1a56765880606db0: Status 404 returned error can't find the container with id de24926ddfd3deb21b4ce351309ca6d639f9da4787bee9bb1a56765880606db0 Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.106812 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod856fc8ab_76a7_4ed4_ab1e_31dad3985982.slice/crio-8f2dcf95584014b60e1df046cf9a0d77fd4a4c80d4324837399d8ffbb7a18513 WatchSource:0}: Error finding container 8f2dcf95584014b60e1df046cf9a0d77fd4a4c80d4324837399d8ffbb7a18513: Status 404 returned error can't find the container with id 8f2dcf95584014b60e1df046cf9a0d77fd4a4c80d4324837399d8ffbb7a18513 Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.109557 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.113056 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1fd35d81_f228_4c45_895a_95feb523ef1f.slice/crio-7c40bd2a77a4afd4ee9f9b83f68acebf32a54e0cda98e83a3c9691d0519de22f WatchSource:0}: Error finding container 7c40bd2a77a4afd4ee9f9b83f68acebf32a54e0cda98e83a3c9691d0519de22f: Status 404 returned error can't find the container with id 7c40bd2a77a4afd4ee9f9b83f68acebf32a54e0cda98e83a3c9691d0519de22f Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.116698 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9x7wv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-748dc6576f-9wnpj_openstack-operators(1e34dcae-a471-4c9c-83f2-43d551516520): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.119889 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9x7wv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-748dc6576f-9wnpj_openstack-operators(1e34dcae-a471-4c9c-83f2-43d551516520): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.121025 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" podUID="1e34dcae-a471-4c9c-83f2-43d551516520" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.229471 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.233298 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1064823e_17fb_42bc_b68e_cfab481fbf8c.slice/crio-f872f854c034fadd6714d2bfca8476b7e77765aebe6b92a903dfe5a6745415ab WatchSource:0}: Error finding container f872f854c034fadd6714d2bfca8476b7e77765aebe6b92a903dfe5a6745415ab: Status 404 returned error can't find the container with id f872f854c034fadd6714d2bfca8476b7e77765aebe6b92a903dfe5a6745415ab Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.235614 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.239550 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83077424_8cf5_48ae_b8b2_d0b94fb650e6.slice/crio-b40564ebb24b3e650145d011ab05d0409511620ba0657c3a5acc08fa2d03eeab WatchSource:0}: Error finding container b40564ebb24b3e650145d011ab05d0409511620ba0657c3a5acc08fa2d03eeab: Status 404 returned error can't find the container with id b40564ebb24b3e650145d011ab05d0409511620ba0657c3a5acc08fa2d03eeab Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.240156 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-hhgb4"] Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.241428 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xkb8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-d64tb_openstack-operators(83077424-8cf5-48ae-b8b2-d0b94fb650e6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.243099 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb75b241_90c4_45ee_9473_864b6b9f8705.slice/crio-7a6a0bed0f2b9756070c8c6940543cd6ffd621e643819c07a6fd7dfaee79f963 WatchSource:0}: Error finding container 7a6a0bed0f2b9756070c8c6940543cd6ffd621e643819c07a6fd7dfaee79f963: Status 404 returned error can't find the container with id 7a6a0bed0f2b9756070c8c6940543cd6ffd621e643819c07a6fd7dfaee79f963 Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.243541 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xkb8c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-6fdc4fcf86-d64tb_openstack-operators(83077424-8cf5-48ae-b8b2-d0b94fb650e6): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.244169 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.244496 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda7c6b74d_a672_4349_b9f7_52a726bc8ec9.slice/crio-27b8e648bf37361544c829670d5e0b8cfcba1b5804f8572d98586c6e691aafce WatchSource:0}: Error finding container 27b8e648bf37361544c829670d5e0b8cfcba1b5804f8572d98586c6e691aafce: Status 404 returned error can't find the container with id 27b8e648bf37361544c829670d5e0b8cfcba1b5804f8572d98586c6e691aafce Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.244632 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" podUID="83077424-8cf5-48ae-b8b2-d0b94fb650e6" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.245669 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fz4fm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-hhgb4_openstack-operators(cb75b241-90c4-45ee-9473-864b6b9f8705): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.246972 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wslns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-gnn9s_openstack-operators(a7c6b74d-a672-4349-b9f7-52a726bc8ec9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.247407 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fz4fm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-hhgb4_openstack-operators(cb75b241-90c4-45ee-9473-864b6b9f8705): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.248683 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" podUID="cb75b241-90c4-45ee-9473-864b6b9f8705" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.249199 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wslns,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-gnn9s_openstack-operators(a7c6b74d-a672-4349-b9f7-52a726bc8ec9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.250928 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" podUID="a7c6b74d-a672-4349-b9f7-52a726bc8ec9" Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.326996 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d60669a_b373_49bf_b65f_88b3e0d48d87.slice/crio-5e5890429c56e26b535fd6b99b2ade2f71045f5b1d7b0883218fcf6426f28a93 WatchSource:0}: Error finding container 5e5890429c56e26b535fd6b99b2ade2f71045f5b1d7b0883218fcf6426f28a93: Status 404 returned error can't find the container with id 5e5890429c56e26b535fd6b99b2ade2f71045f5b1d7b0883218fcf6426f28a93 Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.328026 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66"] Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.329753 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g8mjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-phd66_openstack-operators(7d60669a-b373-49bf-b65f-88b3e0d48d87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.333925 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g8mjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-567f98c9d-phd66_openstack-operators(7d60669a-b373-49bf-b65f-88b3e0d48d87): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.335685 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" podUID="7d60669a-b373-49bf-b65f-88b3e0d48d87" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.336096 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h"] Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.339532 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z6qm7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mc72h_openstack-operators(54a0ed25-ded8-471d-acdd-b34f75347701): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.341282 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" podUID="54a0ed25-ded8-471d-acdd-b34f75347701" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.344998 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96"] Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.352906 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49cb03d2_7430_4f92_b82b_7e4b7ad9faee.slice/crio-cf896a68be7928075f26f5dbd2955d4bf8f43fe3341ce3129d30d9f77b8d2adb WatchSource:0}: Error finding container cf896a68be7928075f26f5dbd2955d4bf8f43fe3341ce3129d30d9f77b8d2adb: Status 404 returned error can't find the container with id cf896a68be7928075f26f5dbd2955d4bf8f43fe3341ce3129d30d9f77b8d2adb Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.353243 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf"] Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.356574 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6w4z4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-d5cc86f4b-j5g96_openstack-operators(49cb03d2-7430-4f92-b82b-7e4b7ad9faee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.358675 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6w4z4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-d5cc86f4b-j5g96_openstack-operators(49cb03d2-7430-4f92-b82b-7e4b7ad9faee): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: W1125 05:52:26.359294 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded82966d_e1bb_448f_be61_091fcb5f3f2b.slice/crio-17350e814b71288bdbfb572052bef0eb7b490e22167a110a8c8b2c3e9e276fe8 WatchSource:0}: Error finding container 17350e814b71288bdbfb572052bef0eb7b490e22167a110a8c8b2c3e9e276fe8: Status 404 returned error can't find the container with id 17350e814b71288bdbfb572052bef0eb7b490e22167a110a8c8b2c3e9e276fe8 Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.360137 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" podUID="49cb03d2-7430-4f92-b82b-7e4b7ad9faee" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.361362 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jhb2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-z9kvf_openstack-operators(ed82966d-e1bb-448f-be61-091fcb5f3f2b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.363137 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jhb2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-z9kvf_openstack-operators(ed82966d-e1bb-448f-be61-091fcb5f3f2b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.364335 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" podUID="ed82966d-e1bb-448f-be61-091fcb5f3f2b" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.636282 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.655117 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3d11487d-d74d-428d-b18b-800238416691-cert\") pod \"openstack-baremetal-operator-controller-manager-b58f89467-gk8ft\" (UID: \"3d11487d-d74d-428d-b18b-800238416691\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.793670 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" event={"ID":"b752b0d3-bffb-4322-9e11-d2a9451c4f39","Type":"ContainerStarted","Data":"401c3228a8c6cb9feab0d7be152bda46164462bf106501394e67ae6e3009e2df"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.795219 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" event={"ID":"cb75b241-90c4-45ee-9473-864b6b9f8705","Type":"ContainerStarted","Data":"7a6a0bed0f2b9756070c8c6940543cd6ffd621e643819c07a6fd7dfaee79f963"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.797138 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" event={"ID":"6af5f14d-1c9e-4985-ba46-3f4b193ac96d","Type":"ContainerStarted","Data":"de24926ddfd3deb21b4ce351309ca6d639f9da4787bee9bb1a56765880606db0"} Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.797496 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" podUID="cb75b241-90c4-45ee-9473-864b6b9f8705" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.798219 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" event={"ID":"dcbf8314-b18c-43ce-8a2e-400b9da62852","Type":"ContainerStarted","Data":"f0598679062135d0d38b26d964049843927b499f02537699ab6ed36da2ea8858"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.800025 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" event={"ID":"83077424-8cf5-48ae-b8b2-d0b94fb650e6","Type":"ContainerStarted","Data":"b40564ebb24b3e650145d011ab05d0409511620ba0657c3a5acc08fa2d03eeab"} Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.801101 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" podUID="83077424-8cf5-48ae-b8b2-d0b94fb650e6" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.801557 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" event={"ID":"856fc8ab-76a7-4ed4-ab1e-31dad3985982","Type":"ContainerStarted","Data":"8f2dcf95584014b60e1df046cf9a0d77fd4a4c80d4324837399d8ffbb7a18513"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.801698 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.802344 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" event={"ID":"54a0ed25-ded8-471d-acdd-b34f75347701","Type":"ContainerStarted","Data":"63ef0bd007fafc5c26d1c5aa2e39808102c7ef6ed7275c2346234e09ae0c98ef"} Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.803140 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" podUID="54a0ed25-ded8-471d-acdd-b34f75347701" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.803495 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" event={"ID":"1fd35d81-f228-4c45-895a-95feb523ef1f","Type":"ContainerStarted","Data":"7c40bd2a77a4afd4ee9f9b83f68acebf32a54e0cda98e83a3c9691d0519de22f"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.804274 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" event={"ID":"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2","Type":"ContainerStarted","Data":"b5d633e17c9c39ee25ff3fdf592ff6b1227f1e818fe153eca1f353f837398cc8"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.806117 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" event={"ID":"ed82966d-e1bb-448f-be61-091fcb5f3f2b","Type":"ContainerStarted","Data":"17350e814b71288bdbfb572052bef0eb7b490e22167a110a8c8b2c3e9e276fe8"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.807461 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" event={"ID":"1064823e-17fb-42bc-b68e-cfab481fbf8c","Type":"ContainerStarted","Data":"f872f854c034fadd6714d2bfca8476b7e77765aebe6b92a903dfe5a6745415ab"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.808317 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" event={"ID":"eef4d5f6-a70a-4b08-8836-3dde6ab0da36","Type":"ContainerStarted","Data":"257fcf90c6decbc07fbf54fd632ea34404e87490c25a9b525857807d6d3d3098"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.810640 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" event={"ID":"49cb03d2-7430-4f92-b82b-7e4b7ad9faee","Type":"ContainerStarted","Data":"cf896a68be7928075f26f5dbd2955d4bf8f43fe3341ce3129d30d9f77b8d2adb"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.817330 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" event={"ID":"17ef75be-981e-4579-ae1b-aa7e4c4b8918","Type":"ContainerStarted","Data":"a30843213d1517b7af7d7a23758bafcfbc74f1b8c95aa16f31f3518ac597c9f5"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.819178 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" event={"ID":"7d60669a-b373-49bf-b65f-88b3e0d48d87","Type":"ContainerStarted","Data":"5e5890429c56e26b535fd6b99b2ade2f71045f5b1d7b0883218fcf6426f28a93"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.822012 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" event={"ID":"b2240efc-4688-4d79-a19d-ea3801742494","Type":"ContainerStarted","Data":"d065e492b8bbde0311975e7f33d60a2b94757414c5d4b07a108b139dd1389fd3"} Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.823347 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" event={"ID":"1e34dcae-a471-4c9c-83f2-43d551516520","Type":"ContainerStarted","Data":"34bf8fe296c2e76bc8caabf32b91790c4807c95d1c85ea6a6546b2f3df8771dd"} Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.825838 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" podUID="1e34dcae-a471-4c9c-83f2-43d551516520" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.825904 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" podUID="49cb03d2-7430-4f92-b82b-7e4b7ad9faee" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.825915 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" podUID="ed82966d-e1bb-448f-be61-091fcb5f3f2b" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.825949 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" podUID="7d60669a-b373-49bf-b65f-88b3e0d48d87" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.829487 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" event={"ID":"a7c6b74d-a672-4349-b9f7-52a726bc8ec9","Type":"ContainerStarted","Data":"27b8e648bf37361544c829670d5e0b8cfcba1b5804f8572d98586c6e691aafce"} Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.834115 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" podUID="a7c6b74d-a672-4349-b9f7-52a726bc8ec9" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.950849 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:26 crc kubenswrapper[4708]: I1125 05:52:26.950971 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.951067 4708 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.951142 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:28.951119943 +0000 UTC m=+690.359953329 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "metrics-server-cert" not found Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.951541 4708 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 05:52:26 crc kubenswrapper[4708]: E1125 05:52:26.951579 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs podName:af33b085-a912-42a9-88ba-6b7e8d27ec8d nodeName:}" failed. No retries permitted until 2025-11-25 05:52:28.951570331 +0000 UTC m=+690.360403718 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs") pod "openstack-operator-controller-manager-7cd5954d9-c7dbb" (UID: "af33b085-a912-42a9-88ba-6b7e8d27ec8d") : secret "webhook-server-cert" not found Nov 25 05:52:27 crc kubenswrapper[4708]: I1125 05:52:27.228178 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft"] Nov 25 05:52:27 crc kubenswrapper[4708]: W1125 05:52:27.243468 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d11487d_d74d_428d_b18b_800238416691.slice/crio-e41edbe30c3026dc44724c1a5b54edbe894b25408ea467ab70b8230b2bea9193 WatchSource:0}: Error finding container e41edbe30c3026dc44724c1a5b54edbe894b25408ea467ab70b8230b2bea9193: Status 404 returned error can't find the container with id e41edbe30c3026dc44724c1a5b54edbe894b25408ea467ab70b8230b2bea9193 Nov 25 05:52:27 crc kubenswrapper[4708]: I1125 05:52:27.838000 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" event={"ID":"3d11487d-d74d-428d-b18b-800238416691","Type":"ContainerStarted","Data":"e41edbe30c3026dc44724c1a5b54edbe894b25408ea467ab70b8230b2bea9193"} Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.839782 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" podUID="54a0ed25-ded8-471d-acdd-b34f75347701" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.840893 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:86df58f744c1d23233cc98f6ea17c8d6da637c50003d0fc8c100045594aa9894\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" podUID="49cb03d2-7430-4f92-b82b-7e4b7ad9faee" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841104 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" podUID="a7c6b74d-a672-4349-b9f7-52a726bc8ec9" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841765 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c0b5f124a37c1538042c0e63f0978429572e2a851d7f3a6eb80de09b86d755a0\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" podUID="83077424-8cf5-48ae-b8b2-d0b94fb650e6" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841832 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" podUID="ed82966d-e1bb-448f-be61-091fcb5f3f2b" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841873 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:5324a6d2f76fc3041023b0cbd09a733ef2b59f310d390e4d6483d219eb96494f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" podUID="7d60669a-b373-49bf-b65f-88b3e0d48d87" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841914 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" podUID="cb75b241-90c4-45ee-9473-864b6b9f8705" Nov 25 05:52:27 crc kubenswrapper[4708]: E1125 05:52:27.841960 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:3ef72bbd7cce89ff54d850ff44ca6d7b2360834a502da3d561aeb6fd3d9af50a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" podUID="1e34dcae-a471-4c9c-83f2-43d551516520" Nov 25 05:52:28 crc kubenswrapper[4708]: I1125 05:52:28.980966 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:28 crc kubenswrapper[4708]: I1125 05:52:28.982577 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:28 crc kubenswrapper[4708]: I1125 05:52:28.987155 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-webhook-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:28 crc kubenswrapper[4708]: I1125 05:52:28.987581 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/af33b085-a912-42a9-88ba-6b7e8d27ec8d-metrics-certs\") pod \"openstack-operator-controller-manager-7cd5954d9-c7dbb\" (UID: \"af33b085-a912-42a9-88ba-6b7e8d27ec8d\") " pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:29 crc kubenswrapper[4708]: I1125 05:52:29.197800 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.640886 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb"] Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.751723 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cxzrz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-b58f89467-gk8ft_openstack-operators(3d11487d-d74d-428d-b18b-800238416691): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.753339 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" podUID="3d11487d-d74d-428d-b18b-800238416691" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.759096 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l2t2l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-fd75fd47d-rmmvr_openstack-operators(6af5f14d-1c9e-4985-ba46-3f4b193ac96d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.763454 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" podUID="6af5f14d-1c9e-4985-ba46-3f4b193ac96d" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.770898 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xwzqt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-7c57c8bbc4-c8c5l_openstack-operators(17ef75be-981e-4579-ae1b-aa7e4c4b8918): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.773700 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" podUID="17ef75be-981e-4579-ae1b-aa7e4c4b8918" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.889843 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" event={"ID":"856fc8ab-76a7-4ed4-ab1e-31dad3985982","Type":"ContainerStarted","Data":"0f599225abe9092a5f51af288156ef091a83d3a744f9577b866698d4590aebec"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.891067 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" event={"ID":"8cb5eab2-d137-4aeb-9ca5-59139c67a2de","Type":"ContainerStarted","Data":"8c7d9c24019ba8b8de8f4b2d8482eeee7f629ea368763126be9e642aa3bfa7cb"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.891970 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" event={"ID":"b2240efc-4688-4d79-a19d-ea3801742494","Type":"ContainerStarted","Data":"eccc32ab6c22515cba26c405d1274838ef03ba3581cd6c626ed66334d9183d2e"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.898044 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" event={"ID":"af33b085-a912-42a9-88ba-6b7e8d27ec8d","Type":"ContainerStarted","Data":"fbed52ca2af6f379559c1f2c5c85f31ef8124ecfa1f80194587614861ce52b49"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.898069 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" event={"ID":"af33b085-a912-42a9-88ba-6b7e8d27ec8d","Type":"ContainerStarted","Data":"9ca49e20aaef5e57a20bdf852fb496793b0fdb8aba8f882810327779a25d4260"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.898696 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.904266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" event={"ID":"fd2a735a-3219-45d8-a689-1a8722923d8d","Type":"ContainerStarted","Data":"cd1a3f9f23b2ea44dff51cd8d7b971c77cb97f50622f1f126fd644894ff8d10e"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.905197 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" event={"ID":"eef4d5f6-a70a-4b08-8836-3dde6ab0da36","Type":"ContainerStarted","Data":"c43f26d97777f98be56c7db364d802d493a7edbff56614a0991935f9b947f9e3"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.909582 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" event={"ID":"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2","Type":"ContainerStarted","Data":"973ed09b0006482ca22f0e00a7e1b5dffcc07067b45e2014fc1028bdecca8a32"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.910836 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" event={"ID":"13a9d474-f528-4913-ada4-313db91a37d5","Type":"ContainerStarted","Data":"196db64d7c24ea52c6071cda30f9aa69014eb76d55222241c3811d0f8834eab6"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.912139 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" event={"ID":"dcbf8314-b18c-43ce-8a2e-400b9da62852","Type":"ContainerStarted","Data":"b59470160499f0c10c375244932be2be1bb93293429bbc0268623e7a5861bdc9"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.913507 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" event={"ID":"17ef75be-981e-4579-ae1b-aa7e4c4b8918","Type":"ContainerStarted","Data":"57e8c757184f2d440d68b3639a833dc479df407ee58d771cab990ce4d269fefb"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.914013 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.914477 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" podUID="17ef75be-981e-4579-ae1b-aa7e4c4b8918" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.915203 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" event={"ID":"3d11487d-d74d-428d-b18b-800238416691","Type":"ContainerStarted","Data":"e1d841874666563d16a91dd8c902017352236599c876c2868c4f677a7d5dcefa"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.915628 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.916082 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" podUID="3d11487d-d74d-428d-b18b-800238416691" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.917270 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" event={"ID":"6af5f14d-1c9e-4985-ba46-3f4b193ac96d","Type":"ContainerStarted","Data":"e843f8fd695725ebf2a98cffb92c13f590f930e8a0a547d7e46576240e4b364f"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.917665 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:34 crc kubenswrapper[4708]: E1125 05:52:34.918020 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" podUID="6af5f14d-1c9e-4985-ba46-3f4b193ac96d" Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.919075 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" event={"ID":"1064823e-17fb-42bc-b68e-cfab481fbf8c","Type":"ContainerStarted","Data":"80feb8249dd03b0d180bc4f71c87387cc2077b7a2c668dfbc8b76026ddfc88e3"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.920202 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" event={"ID":"1fd35d81-f228-4c45-895a-95feb523ef1f","Type":"ContainerStarted","Data":"3ed10765517bdd6292762997e8e1e285239057921bc33a156baf25e4d43790dc"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.922661 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" event={"ID":"b752b0d3-bffb-4322-9e11-d2a9451c4f39","Type":"ContainerStarted","Data":"84edf3f8a5caf79ec8bf6c93715081ae2cef0175e95e242665d929f704f6a3e2"} Nov 25 05:52:34 crc kubenswrapper[4708]: I1125 05:52:34.958055 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" podStartSLOduration=9.958044427 podStartE2EDuration="9.958044427s" podCreationTimestamp="2025-11-25 05:52:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:52:34.957666976 +0000 UTC m=+696.366500363" watchObservedRunningTime="2025-11-25 05:52:34.958044427 +0000 UTC m=+696.366877813" Nov 25 05:52:35 crc kubenswrapper[4708]: E1125 05:52:35.945779 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" podUID="6af5f14d-1c9e-4985-ba46-3f4b193ac96d" Nov 25 05:52:35 crc kubenswrapper[4708]: E1125 05:52:35.946018 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" podUID="17ef75be-981e-4579-ae1b-aa7e4c4b8918" Nov 25 05:52:35 crc kubenswrapper[4708]: E1125 05:52:35.946290 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" podUID="3d11487d-d74d-428d-b18b-800238416691" Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.972986 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" event={"ID":"b752b0d3-bffb-4322-9e11-d2a9451c4f39","Type":"ContainerStarted","Data":"682f3127a00e27769aaaa0eba834e2cd083223870107d53039cf66eabb112da9"} Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.973578 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.983707 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" event={"ID":"856fc8ab-76a7-4ed4-ab1e-31dad3985982","Type":"ContainerStarted","Data":"c953c43aca04c15cdc824a25c30778781c0f3ce8371e5afa5c4904a38f54eef6"} Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.984405 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.997663 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" event={"ID":"1064823e-17fb-42bc-b68e-cfab481fbf8c","Type":"ContainerStarted","Data":"8a27f3ba122c051cd589a620ca67a374a37946b67c996849902725637b1ce694"} Nov 25 05:52:37 crc kubenswrapper[4708]: I1125 05:52:37.998123 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.000143 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" event={"ID":"eef4d5f6-a70a-4b08-8836-3dde6ab0da36","Type":"ContainerStarted","Data":"fc49d0a067fbb269d9d6bfbafed32d86a2e1c9390ffb96cc593756eaa04187c1"} Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.000792 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.001608 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" podStartSLOduration=2.740786327 podStartE2EDuration="14.001598366s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.888197037 +0000 UTC m=+687.297030424" lastFinishedPulling="2025-11-25 05:52:37.149009077 +0000 UTC m=+698.557842463" observedRunningTime="2025-11-25 05:52:37.999403792 +0000 UTC m=+699.408237179" watchObservedRunningTime="2025-11-25 05:52:38.001598366 +0000 UTC m=+699.410431753" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.020103 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" event={"ID":"8cb5eab2-d137-4aeb-9ca5-59139c67a2de","Type":"ContainerStarted","Data":"463ca71f3fa2f413bf2b431b658e5416ae9c2c049fb844dc4f48699c77d312be"} Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.020376 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.030730 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" event={"ID":"b2240efc-4688-4d79-a19d-ea3801742494","Type":"ContainerStarted","Data":"71d0d2d491da9e29f562af765ea94b9e3371951a737a7fda50e410a6461ea440"} Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.031256 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.034734 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" podStartSLOduration=2.682492947 podStartE2EDuration="14.034715604s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.110958203 +0000 UTC m=+687.519791590" lastFinishedPulling="2025-11-25 05:52:37.463180861 +0000 UTC m=+698.872014247" observedRunningTime="2025-11-25 05:52:38.030549464 +0000 UTC m=+699.439382851" watchObservedRunningTime="2025-11-25 05:52:38.034715604 +0000 UTC m=+699.443548989" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.058034 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" event={"ID":"dcbf8314-b18c-43ce-8a2e-400b9da62852","Type":"ContainerStarted","Data":"fda16050b1a0ec3536295bde114c21bb3b74bb36b09f10dbf4d876a43458dddc"} Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.058886 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.060704 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" podStartSLOduration=2.529669148 podStartE2EDuration="14.060692629s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.23655931 +0000 UTC m=+687.645392696" lastFinishedPulling="2025-11-25 05:52:37.767582791 +0000 UTC m=+699.176416177" observedRunningTime="2025-11-25 05:52:38.05907965 +0000 UTC m=+699.467913036" watchObservedRunningTime="2025-11-25 05:52:38.060692629 +0000 UTC m=+699.469526015" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.093057 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" podStartSLOduration=2.495573087 podStartE2EDuration="14.093038641s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.819382225 +0000 UTC m=+687.228215611" lastFinishedPulling="2025-11-25 05:52:37.416847779 +0000 UTC m=+698.825681165" observedRunningTime="2025-11-25 05:52:38.083793055 +0000 UTC m=+699.492626441" watchObservedRunningTime="2025-11-25 05:52:38.093038641 +0000 UTC m=+699.501872027" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.115732 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" podStartSLOduration=2.55518877 podStartE2EDuration="14.115712414s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.800914115 +0000 UTC m=+687.209747500" lastFinishedPulling="2025-11-25 05:52:37.361437758 +0000 UTC m=+698.770271144" observedRunningTime="2025-11-25 05:52:38.109388482 +0000 UTC m=+699.518221868" watchObservedRunningTime="2025-11-25 05:52:38.115712414 +0000 UTC m=+699.524545800" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.130856 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" podStartSLOduration=2.181095668 podStartE2EDuration="14.130840552s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.624270766 +0000 UTC m=+687.033104153" lastFinishedPulling="2025-11-25 05:52:37.57401565 +0000 UTC m=+698.982849037" observedRunningTime="2025-11-25 05:52:38.127615266 +0000 UTC m=+699.536448643" watchObservedRunningTime="2025-11-25 05:52:38.130840552 +0000 UTC m=+699.539673938" Nov 25 05:52:38 crc kubenswrapper[4708]: I1125 05:52:38.147908 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" podStartSLOduration=3.0899625139999998 podStartE2EDuration="14.147894098s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.100581405 +0000 UTC m=+687.509414782" lastFinishedPulling="2025-11-25 05:52:37.158512981 +0000 UTC m=+698.567346366" observedRunningTime="2025-11-25 05:52:38.145259405 +0000 UTC m=+699.554092790" watchObservedRunningTime="2025-11-25 05:52:38.147894098 +0000 UTC m=+699.556727484" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.070123 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" event={"ID":"1fd35d81-f228-4c45-895a-95feb523ef1f","Type":"ContainerStarted","Data":"4dfdfa9000188f12872acd36c4cdfd6e1117f349c7f0637d18cfc25b858d604b"} Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.070589 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.074096 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.075012 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" event={"ID":"e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2","Type":"ContainerStarted","Data":"078ffa43de222ac7f6fa3f6fc0d7e7bc60ad550cf2534e68edd2974a9f6ca0be"} Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.075547 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.077481 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" event={"ID":"13a9d474-f528-4913-ada4-313db91a37d5","Type":"ContainerStarted","Data":"766375c59797b186c9abe3dc2767a11d915a228efb8351f666bcc5c5b13bf974"} Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.077678 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.077927 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.080661 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.080982 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" event={"ID":"fd2a735a-3219-45d8-a689-1a8722923d8d","Type":"ContainerStarted","Data":"02031e51d3760767c59b062aa9339b26df73ccd6e1fe5732ee9327fe7c516f33"} Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.084587 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-qjcbh" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.084675 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-kvrrg" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.084913 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-vzz64" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.085162 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-x4xzv" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.085211 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-6zc24" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.085755 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-68b95954c9-df8zr" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.085835 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-j98rt" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.093423 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-8xsh6" podStartSLOduration=3.342416002 podStartE2EDuration="15.093406986s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.116101622 +0000 UTC m=+687.524935008" lastFinishedPulling="2025-11-25 05:52:37.867092606 +0000 UTC m=+699.275925992" observedRunningTime="2025-11-25 05:52:39.088337327 +0000 UTC m=+700.497170702" watchObservedRunningTime="2025-11-25 05:52:39.093406986 +0000 UTC m=+700.502240372" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.214666 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" podStartSLOduration=3.065624486 podStartE2EDuration="15.214641267s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.692395008 +0000 UTC m=+687.101228384" lastFinishedPulling="2025-11-25 05:52:37.84141178 +0000 UTC m=+699.250245165" observedRunningTime="2025-11-25 05:52:39.182632909 +0000 UTC m=+700.591466295" watchObservedRunningTime="2025-11-25 05:52:39.214641267 +0000 UTC m=+700.623474653" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.224627 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7cd5954d9-c7dbb" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.250044 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-9knwb" podStartSLOduration=2.831055086 podStartE2EDuration="15.250021195s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.862116757 +0000 UTC m=+687.270950144" lastFinishedPulling="2025-11-25 05:52:38.281082867 +0000 UTC m=+699.689916253" observedRunningTime="2025-11-25 05:52:39.242415819 +0000 UTC m=+700.651249204" watchObservedRunningTime="2025-11-25 05:52:39.250021195 +0000 UTC m=+700.658854582" Nov 25 05:52:39 crc kubenswrapper[4708]: I1125 05:52:39.287782 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-b7hd2" podStartSLOduration=3.02910045 podStartE2EDuration="15.287747072s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:25.625855782 +0000 UTC m=+687.034689167" lastFinishedPulling="2025-11-25 05:52:37.884502403 +0000 UTC m=+699.293335789" observedRunningTime="2025-11-25 05:52:39.284218105 +0000 UTC m=+700.693051492" watchObservedRunningTime="2025-11-25 05:52:39.287747072 +0000 UTC m=+700.696580459" Nov 25 05:52:40 crc kubenswrapper[4708]: I1125 05:52:40.091029 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:40 crc kubenswrapper[4708]: I1125 05:52:40.092711 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-j2vk8" Nov 25 05:52:44 crc kubenswrapper[4708]: I1125 05:52:44.357365 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:52:44 crc kubenswrapper[4708]: I1125 05:52:44.357883 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.107656 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.152817 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" event={"ID":"83077424-8cf5-48ae-b8b2-d0b94fb650e6","Type":"ContainerStarted","Data":"6643e7a5f009b151dcfe37f80d9e00e007ded3d62478927ce2591376cea28969"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.152856 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" event={"ID":"83077424-8cf5-48ae-b8b2-d0b94fb650e6","Type":"ContainerStarted","Data":"eabfda60e776f1559c61daddeab2d138dfcc5283b9a51d85b5e92d510ebc397f"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.153482 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.156855 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" event={"ID":"cb75b241-90c4-45ee-9473-864b6b9f8705","Type":"ContainerStarted","Data":"6098602291e956eff783d6e530c09ed47a49a95b46e52b505b713101cd76efc2"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.156898 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" event={"ID":"cb75b241-90c4-45ee-9473-864b6b9f8705","Type":"ContainerStarted","Data":"6439f0b9e3a50e4026fbb2016a9abaa3cfc98e2264e7c6ea145eece54afce3cf"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.157058 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.158833 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" event={"ID":"1e34dcae-a471-4c9c-83f2-43d551516520","Type":"ContainerStarted","Data":"c389df2d7792082b3864495d5b50e7972e1d5267f2c15d2787e81c4694ed03da"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.158861 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" event={"ID":"1e34dcae-a471-4c9c-83f2-43d551516520","Type":"ContainerStarted","Data":"e88f5252087cc35adea55b3464f3ca59bc53cb861fcafc322864301ece90f3d8"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.159181 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.161377 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" event={"ID":"7d60669a-b373-49bf-b65f-88b3e0d48d87","Type":"ContainerStarted","Data":"eb25de5ffc5902909fdd12468533026319987eca4e33198c7deb8499f99881d4"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.161403 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" event={"ID":"7d60669a-b373-49bf-b65f-88b3e0d48d87","Type":"ContainerStarted","Data":"b6c9d916b0ec424069e726fba3f915c61afefe57c3ba737d0e9340e150afd6a7"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.161785 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.163967 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" event={"ID":"ed82966d-e1bb-448f-be61-091fcb5f3f2b","Type":"ContainerStarted","Data":"75045239ba93269d308e916b0282f1cd76ecfbb2dbf34b10921f66128aff390e"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.163992 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" event={"ID":"ed82966d-e1bb-448f-be61-091fcb5f3f2b","Type":"ContainerStarted","Data":"d323bc5cfbac448e30a99dcf50b6ddd1d59e6e80226151407b5c20134a1960ae"} Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.164319 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.171480 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" podStartSLOduration=2.98542703 podStartE2EDuration="21.171469197s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.24130041 +0000 UTC m=+687.650133787" lastFinishedPulling="2025-11-25 05:52:44.427342567 +0000 UTC m=+705.836175954" observedRunningTime="2025-11-25 05:52:45.16745744 +0000 UTC m=+706.576290825" watchObservedRunningTime="2025-11-25 05:52:45.171469197 +0000 UTC m=+706.580302573" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.191775 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" podStartSLOduration=1.9884214519999999 podStartE2EDuration="20.191755472s" podCreationTimestamp="2025-11-25 05:52:25 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.245405334 +0000 UTC m=+687.654238720" lastFinishedPulling="2025-11-25 05:52:44.448739354 +0000 UTC m=+705.857572740" observedRunningTime="2025-11-25 05:52:45.191577456 +0000 UTC m=+706.600410842" watchObservedRunningTime="2025-11-25 05:52:45.191755472 +0000 UTC m=+706.600588858" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.222379 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" podStartSLOduration=2.160953855 podStartE2EDuration="20.222366256s" podCreationTimestamp="2025-11-25 05:52:25 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.361284406 +0000 UTC m=+687.770117792" lastFinishedPulling="2025-11-25 05:52:44.422696806 +0000 UTC m=+705.831530193" observedRunningTime="2025-11-25 05:52:45.219142713 +0000 UTC m=+706.627976099" watchObservedRunningTime="2025-11-25 05:52:45.222366256 +0000 UTC m=+706.631199632" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.234462 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" podStartSLOduration=3.111205953 podStartE2EDuration="21.234455571s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.329553982 +0000 UTC m=+687.738387369" lastFinishedPulling="2025-11-25 05:52:44.452803601 +0000 UTC m=+705.861636987" observedRunningTime="2025-11-25 05:52:45.232839706 +0000 UTC m=+706.641673093" watchObservedRunningTime="2025-11-25 05:52:45.234455571 +0000 UTC m=+706.643288956" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.248911 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" podStartSLOduration=2.943341921 podStartE2EDuration="21.248890332s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.116423168 +0000 UTC m=+687.525256554" lastFinishedPulling="2025-11-25 05:52:44.42197158 +0000 UTC m=+705.830804965" observedRunningTime="2025-11-25 05:52:45.247064071 +0000 UTC m=+706.655897457" watchObservedRunningTime="2025-11-25 05:52:45.248890332 +0000 UTC m=+706.657723718" Nov 25 05:52:45 crc kubenswrapper[4708]: I1125 05:52:45.262857 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" Nov 25 05:52:46 crc kubenswrapper[4708]: I1125 05:52:46.812451 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.191081 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" event={"ID":"a7c6b74d-a672-4349-b9f7-52a726bc8ec9","Type":"ContainerStarted","Data":"661d918ab2d87784f03931c6caf6aadc2d1e5bbaa89cd5a84b24826aa81e6b90"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.191393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" event={"ID":"a7c6b74d-a672-4349-b9f7-52a726bc8ec9","Type":"ContainerStarted","Data":"efc81d1e007ec4d1507932edf6680ff697a2413590067245bc87a02fe33d5111"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.191574 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.192963 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" event={"ID":"6af5f14d-1c9e-4985-ba46-3f4b193ac96d","Type":"ContainerStarted","Data":"6bb32b044954175f144d6b379ded9558e45af49371e20ed0fb5de01db5266032"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.195211 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" event={"ID":"17ef75be-981e-4579-ae1b-aa7e4c4b8918","Type":"ContainerStarted","Data":"52ab58535dfe45410558a0c337a5be818306067a63d15107173774f04879c9fb"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.196659 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" event={"ID":"54a0ed25-ded8-471d-acdd-b34f75347701","Type":"ContainerStarted","Data":"60b8c15927c56b6c2b58a1c905b5be8c7fd363e165c8ef908cea86470597f304"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.199268 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" event={"ID":"3d11487d-d74d-428d-b18b-800238416691","Type":"ContainerStarted","Data":"eee88a9913c1496b229c8cb1e0ac167d75e31f398fd6e2d46ec1897c0f7d0f08"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.201611 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" event={"ID":"49cb03d2-7430-4f92-b82b-7e4b7ad9faee","Type":"ContainerStarted","Data":"685b15c39e30637f1eec6ccbaaf127c1adf06b065d4462cdd9ec72e14b4af374"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.201644 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" event={"ID":"49cb03d2-7430-4f92-b82b-7e4b7ad9faee","Type":"ContainerStarted","Data":"33e9052418c10545da8a23b488ead6113584f9beea545c6832014fe8ba54365c"} Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.202070 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.213500 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" podStartSLOduration=3.057249388 podStartE2EDuration="24.213487353s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.246848032 +0000 UTC m=+687.655681418" lastFinishedPulling="2025-11-25 05:52:47.403085997 +0000 UTC m=+708.811919383" observedRunningTime="2025-11-25 05:52:48.21265178 +0000 UTC m=+709.621485166" watchObservedRunningTime="2025-11-25 05:52:48.213487353 +0000 UTC m=+709.622320740" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.241929 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-b58f89467-gk8ft" podStartSLOduration=17.208353034 podStartE2EDuration="24.24191711s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:27.246439212 +0000 UTC m=+688.655272599" lastFinishedPulling="2025-11-25 05:52:34.280003289 +0000 UTC m=+695.688836675" observedRunningTime="2025-11-25 05:52:48.235092443 +0000 UTC m=+709.643925829" watchObservedRunningTime="2025-11-25 05:52:48.24191711 +0000 UTC m=+709.650750496" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.264669 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" podStartSLOduration=3.221728713 podStartE2EDuration="24.264649052s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.356343217 +0000 UTC m=+687.765176604" lastFinishedPulling="2025-11-25 05:52:47.399263556 +0000 UTC m=+708.808096943" observedRunningTime="2025-11-25 05:52:48.256481797 +0000 UTC m=+709.665315173" watchObservedRunningTime="2025-11-25 05:52:48.264649052 +0000 UTC m=+709.673482438" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.274281 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mc72h" podStartSLOduration=2.195249511 podStartE2EDuration="23.274267521s" podCreationTimestamp="2025-11-25 05:52:25 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.339371537 +0000 UTC m=+687.748204922" lastFinishedPulling="2025-11-25 05:52:47.418389545 +0000 UTC m=+708.827222932" observedRunningTime="2025-11-25 05:52:48.270466381 +0000 UTC m=+709.679299766" watchObservedRunningTime="2025-11-25 05:52:48.274267521 +0000 UTC m=+709.683100907" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.286154 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-rmmvr" podStartSLOduration=16.080494804 podStartE2EDuration="24.286131791s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.082985388 +0000 UTC m=+687.491818774" lastFinishedPulling="2025-11-25 05:52:34.288622386 +0000 UTC m=+695.697455761" observedRunningTime="2025-11-25 05:52:48.284406522 +0000 UTC m=+709.693239907" watchObservedRunningTime="2025-11-25 05:52:48.286131791 +0000 UTC m=+709.694965177" Nov 25 05:52:48 crc kubenswrapper[4708]: I1125 05:52:48.301846 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-c8c5l" podStartSLOduration=16.092585278 podStartE2EDuration="24.301812129s" podCreationTimestamp="2025-11-25 05:52:24 +0000 UTC" firstStartedPulling="2025-11-25 05:52:26.068073397 +0000 UTC m=+687.476906773" lastFinishedPulling="2025-11-25 05:52:34.277300237 +0000 UTC m=+695.686133624" observedRunningTime="2025-11-25 05:52:48.297617588 +0000 UTC m=+709.706450974" watchObservedRunningTime="2025-11-25 05:52:48.301812129 +0000 UTC m=+709.710645506" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.077638 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-9wnpj" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.325612 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-gnn9s" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.378566 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-d64tb" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.389716 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-phd66" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.450051 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-z9kvf" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.469812 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-hhgb4" Nov 25 05:52:55 crc kubenswrapper[4708]: I1125 05:52:55.589652 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-d5cc86f4b-j5g96" Nov 25 05:53:02 crc kubenswrapper[4708]: I1125 05:53:02.096056 4708 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.742031 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.744658 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.748116 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.748419 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-fk9r4" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.748259 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.748683 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.753898 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.788426 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.790102 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.792655 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.807175 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.923428 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.923488 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.923532 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df6f8\" (UniqueName: \"kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.923561 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6dvw\" (UniqueName: \"kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:10 crc kubenswrapper[4708]: I1125 05:53:10.924039 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.025611 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df6f8\" (UniqueName: \"kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.025672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6dvw\" (UniqueName: \"kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.025798 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.025921 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.026053 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.027086 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.027134 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.027182 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.044761 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6dvw\" (UniqueName: \"kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw\") pod \"dnsmasq-dns-f48db6df-z84xs\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.044821 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df6f8\" (UniqueName: \"kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8\") pod \"dnsmasq-dns-56fbcfc7-dhtdd\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.061598 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.108472 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.472888 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:11 crc kubenswrapper[4708]: W1125 05:53:11.539415 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce3c89d2_03cc_4a71_ae7e_d49a930c9b39.slice/crio-bf000bf6560685b48687d24220a4d2b602dde428914039210413a5142f181c5e WatchSource:0}: Error finding container bf000bf6560685b48687d24220a4d2b602dde428914039210413a5142f181c5e: Status 404 returned error can't find the container with id bf000bf6560685b48687d24220a4d2b602dde428914039210413a5142f181c5e Nov 25 05:53:11 crc kubenswrapper[4708]: I1125 05:53:11.540755 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:12 crc kubenswrapper[4708]: I1125 05:53:12.392804 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f48db6df-z84xs" event={"ID":"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39","Type":"ContainerStarted","Data":"bf000bf6560685b48687d24220a4d2b602dde428914039210413a5142f181c5e"} Nov 25 05:53:12 crc kubenswrapper[4708]: I1125 05:53:12.394677 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" event={"ID":"e291e86e-e697-4d41-8e2d-495f975371f0","Type":"ContainerStarted","Data":"2e4d873b23e244cef331a9f18bd12cba509552822fe2b628a2e72070e7e051b2"} Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.270794 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.304605 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.354590 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.374067 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.474886 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdw2q\" (UniqueName: \"kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.475469 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.475571 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.568929 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.577843 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdw2q\" (UniqueName: \"kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.577894 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.577923 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.578777 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.579555 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.593207 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.594418 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.597169 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdw2q\" (UniqueName: \"kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q\") pod \"dnsmasq-dns-6c965d56d7-dkqw7\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.604767 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.733678 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.782919 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.783149 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdn2h\" (UniqueName: \"kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.783370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.884538 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.884619 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdn2h\" (UniqueName: \"kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.884680 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.885453 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.885532 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.911849 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdn2h\" (UniqueName: \"kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h\") pod \"dnsmasq-dns-7bd6d59bf-n6v9f\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:13 crc kubenswrapper[4708]: I1125 05:53:13.933856 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.129688 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:14 crc kubenswrapper[4708]: W1125 05:53:14.136915 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2ef49314_e8b4_4aee_8a73_e10b2dc6e461.slice/crio-e26ef4c662c1c8cff78185824b9171d794c9ebcace3433d1461b89876ca9d2cf WatchSource:0}: Error finding container e26ef4c662c1c8cff78185824b9171d794c9ebcace3433d1461b89876ca9d2cf: Status 404 returned error can't find the container with id e26ef4c662c1c8cff78185824b9171d794c9ebcace3433d1461b89876ca9d2cf Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.342431 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.357278 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.357380 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.442893 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" event={"ID":"2ef49314-e8b4-4aee-8a73-e10b2dc6e461","Type":"ContainerStarted","Data":"e26ef4c662c1c8cff78185824b9171d794c9ebcace3433d1461b89876ca9d2cf"} Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.448979 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" event={"ID":"3226b19a-018d-47fd-8118-abaa87271f85","Type":"ContainerStarted","Data":"9ce7494b2985534bde9195b8e1f7baeb2a58563f53b3dcf2d5168310c35719f1"} Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.452366 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.453918 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.457224 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.457604 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.457799 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.457935 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-l9958" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.458230 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.459615 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.459989 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.464987 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.597616 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.597681 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.597716 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.597850 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.597928 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtkqd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598131 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598187 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598420 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598463 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.598494 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.700813 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.700922 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701020 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701106 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701136 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701194 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701281 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.701269 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.702583 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.702643 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.702751 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.702844 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtkqd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.702930 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.705012 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.705538 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.705634 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.706867 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.708961 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.710560 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.710765 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.719088 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.728301 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.729446 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.743898 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.744192 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.744235 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.745575 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.746029 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.746216 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.746396 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.746537 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-brbzt" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.750490 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtkqd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.753999 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.784160 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.907674 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.907935 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908000 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908137 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908359 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908420 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8s4h\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908639 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908774 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908802 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908857 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:14 crc kubenswrapper[4708]: I1125 05:53:14.908942 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.011708 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012274 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012301 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012383 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012474 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012509 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.012603 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8s4h\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013269 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013277 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013276 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013501 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013501 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013565 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.013605 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.015087 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.015180 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.015877 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.017818 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.019373 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.020296 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.020963 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.033150 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8s4h\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.039165 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.101142 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.191163 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.456587 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerStarted","Data":"0aab5c063b1fed19ecdce5f3bf0c14b847ba6f45acbbdf3e061db1e0830547e6"} Nov 25 05:53:15 crc kubenswrapper[4708]: I1125 05:53:15.551778 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:53:15 crc kubenswrapper[4708]: W1125 05:53:15.562715 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9fe6604_dc04_4655_9a26_53faf3d3b6fa.slice/crio-aeb47cd0b95259e830debc44164147250dc98d3e28dcf9b755bf19d156b4578c WatchSource:0}: Error finding container aeb47cd0b95259e830debc44164147250dc98d3e28dcf9b755bf19d156b4578c: Status 404 returned error can't find the container with id aeb47cd0b95259e830debc44164147250dc98d3e28dcf9b755bf19d156b4578c Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.178077 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.179697 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.185126 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.186175 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-mmdcw" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.186411 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.186793 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.186911 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.188181 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335463 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335512 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335572 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kube-api-access-tp6sz\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335595 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335632 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-default\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335695 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335716 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.335777 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kolla-config\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.436995 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437046 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kolla-config\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437129 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437158 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437178 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kube-api-access-tp6sz\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437198 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437218 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-default\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437340 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437696 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.437706 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-generated\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.438713 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kolla-config\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.440056 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-operator-scripts\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.440298 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-config-data-default\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.445908 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.447462 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.465719 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp6sz\" (UniqueName: \"kubernetes.io/projected/33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc-kube-api-access-tp6sz\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.472233 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc\") " pod="openstack/openstack-galera-0" Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.486726 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerStarted","Data":"aeb47cd0b95259e830debc44164147250dc98d3e28dcf9b755bf19d156b4578c"} Nov 25 05:53:16 crc kubenswrapper[4708]: I1125 05:53:16.495299 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.019554 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 05:53:17 crc kubenswrapper[4708]: W1125 05:53:17.028335 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod33d2c4ee_3cf2_434f_bbaf_b2087a17b0bc.slice/crio-051fb7a3344a43eb0ee85b5e698aa949da59da63bc133e3f222000190beb33c7 WatchSource:0}: Error finding container 051fb7a3344a43eb0ee85b5e698aa949da59da63bc133e3f222000190beb33c7: Status 404 returned error can't find the container with id 051fb7a3344a43eb0ee85b5e698aa949da59da63bc133e3f222000190beb33c7 Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.505987 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc","Type":"ContainerStarted","Data":"051fb7a3344a43eb0ee85b5e698aa949da59da63bc133e3f222000190beb33c7"} Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.702014 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.704828 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.710231 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.710439 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-xqzdd" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.710502 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.710714 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.724364 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.803125 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.813747 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.816758 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.820328 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.824219 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-xgtf4" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.860354 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866772 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866825 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866880 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b8379446-127e-4f18-b730-085e4b80e60b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866910 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnwc2\" (UniqueName: \"kubernetes.io/projected/b8379446-127e-4f18-b730-085e4b80e60b-kube-api-access-wnwc2\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866962 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.866985 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.867004 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.867021 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969000 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-config-data\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969052 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kolla-config\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969095 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969114 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969170 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969199 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969247 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969298 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttz7h\" (UniqueName: \"kubernetes.io/projected/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kube-api-access-ttz7h\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969343 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969371 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.969938 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.970063 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.972098 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.972185 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b8379446-127e-4f18-b730-085e4b80e60b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.972242 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnwc2\" (UniqueName: \"kubernetes.io/projected/b8379446-127e-4f18-b730-085e4b80e60b-kube-api-access-wnwc2\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.972545 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/b8379446-127e-4f18-b730-085e4b80e60b-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.977835 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.978508 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/b8379446-127e-4f18-b730-085e4b80e60b-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.987168 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.990478 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnwc2\" (UniqueName: \"kubernetes.io/projected/b8379446-127e-4f18-b730-085e4b80e60b-kube-api-access-wnwc2\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:17 crc kubenswrapper[4708]: I1125 05:53:17.995703 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.003965 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8379446-127e-4f18-b730-085e4b80e60b-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"b8379446-127e-4f18-b730-085e4b80e60b\") " pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.044494 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.074747 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.074842 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.074927 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-config-data\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.074956 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kolla-config\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.075619 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttz7h\" (UniqueName: \"kubernetes.io/projected/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kube-api-access-ttz7h\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.075797 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-config-data\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.075834 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kolla-config\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.078623 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.079753 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.089959 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttz7h\" (UniqueName: \"kubernetes.io/projected/c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8-kube-api-access-ttz7h\") pod \"memcached-0\" (UID: \"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8\") " pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.133858 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.515006 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 05:53:18 crc kubenswrapper[4708]: W1125 05:53:18.523162 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8379446_127e_4f18_b730_085e4b80e60b.slice/crio-54c685943dc1220b9be21e149535e112d2dbec35e12a1ce450b8160aca327dec WatchSource:0}: Error finding container 54c685943dc1220b9be21e149535e112d2dbec35e12a1ce450b8160aca327dec: Status 404 returned error can't find the container with id 54c685943dc1220b9be21e149535e112d2dbec35e12a1ce450b8160aca327dec Nov 25 05:53:18 crc kubenswrapper[4708]: I1125 05:53:18.641598 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 05:53:18 crc kubenswrapper[4708]: W1125 05:53:18.648271 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc2a8af8f_ba8a_43c0_84dd_d3d7f1fceaf8.slice/crio-12672527aaeb3b9848b0ae6dcf7e6721f73cf68ec94cb91bab80bfd8c9087a98 WatchSource:0}: Error finding container 12672527aaeb3b9848b0ae6dcf7e6721f73cf68ec94cb91bab80bfd8c9087a98: Status 404 returned error can't find the container with id 12672527aaeb3b9848b0ae6dcf7e6721f73cf68ec94cb91bab80bfd8c9087a98 Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.527320 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8","Type":"ContainerStarted","Data":"12672527aaeb3b9848b0ae6dcf7e6721f73cf68ec94cb91bab80bfd8c9087a98"} Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.529210 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b8379446-127e-4f18-b730-085e4b80e60b","Type":"ContainerStarted","Data":"54c685943dc1220b9be21e149535e112d2dbec35e12a1ce450b8160aca327dec"} Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.703086 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.704071 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.707884 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-k6bw7" Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.713077 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dw7x\" (UniqueName: \"kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x\") pod \"kube-state-metrics-0\" (UID: \"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460\") " pod="openstack/kube-state-metrics-0" Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.727280 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.815023 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dw7x\" (UniqueName: \"kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x\") pod \"kube-state-metrics-0\" (UID: \"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460\") " pod="openstack/kube-state-metrics-0" Nov 25 05:53:19 crc kubenswrapper[4708]: I1125 05:53:19.837655 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dw7x\" (UniqueName: \"kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x\") pod \"kube-state-metrics-0\" (UID: \"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460\") " pod="openstack/kube-state-metrics-0" Nov 25 05:53:20 crc kubenswrapper[4708]: I1125 05:53:20.036594 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:53:20 crc kubenswrapper[4708]: I1125 05:53:20.488490 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:53:20 crc kubenswrapper[4708]: W1125 05:53:20.492887 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0dbf1bb3_7a60_4c3e_a942_cab2c9a57460.slice/crio-ced89f359db24b22d05f83622be3d9d37d9d23bff403a31a69e0b72efbcb67f8 WatchSource:0}: Error finding container ced89f359db24b22d05f83622be3d9d37d9d23bff403a31a69e0b72efbcb67f8: Status 404 returned error can't find the container with id ced89f359db24b22d05f83622be3d9d37d9d23bff403a31a69e0b72efbcb67f8 Nov 25 05:53:20 crc kubenswrapper[4708]: I1125 05:53:20.538779 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460","Type":"ContainerStarted","Data":"ced89f359db24b22d05f83622be3d9d37d9d23bff403a31a69e0b72efbcb67f8"} Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.582847 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460","Type":"ContainerStarted","Data":"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452"} Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.583321 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.605461 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.608548771 podStartE2EDuration="4.605442439s" podCreationTimestamp="2025-11-25 05:53:19 +0000 UTC" firstStartedPulling="2025-11-25 05:53:20.495939037 +0000 UTC m=+741.904772423" lastFinishedPulling="2025-11-25 05:53:22.492832705 +0000 UTC m=+743.901666091" observedRunningTime="2025-11-25 05:53:23.602673824 +0000 UTC m=+745.011507210" watchObservedRunningTime="2025-11-25 05:53:23.605442439 +0000 UTC m=+745.014275826" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.723502 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-qc989"] Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.724566 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.726738 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.727049 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-w42nw" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.727161 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.730576 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-xnwzt"] Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.732532 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.738594 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qc989"] Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.750860 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xnwzt"] Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788202 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg7f4\" (UniqueName: \"kubernetes.io/projected/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-kube-api-access-wg7f4\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788270 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pztq\" (UniqueName: \"kubernetes.io/projected/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-kube-api-access-5pztq\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788295 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788323 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-log-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788343 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788363 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-combined-ca-bundle\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788384 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-lib\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788405 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-scripts\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788425 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-scripts\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788451 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-etc-ovs\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788477 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-ovn-controller-tls-certs\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788497 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-log\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.788547 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-run\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894652 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-scripts\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894711 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-scripts\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894751 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-etc-ovs\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894798 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-ovn-controller-tls-certs\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894830 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-log\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894868 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-run\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894889 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg7f4\" (UniqueName: \"kubernetes.io/projected/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-kube-api-access-wg7f4\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894916 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pztq\" (UniqueName: \"kubernetes.io/projected/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-kube-api-access-5pztq\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894942 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894973 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-log-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.894998 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.895019 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-combined-ca-bundle\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.895050 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-lib\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.895912 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-etc-ovs\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.896127 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-run\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.896369 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-log\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.896486 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-log-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.896560 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.896654 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-var-run-ovn\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.899577 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-scripts\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.900181 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-var-lib\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.900418 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-scripts\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.908812 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-ovn-controller-tls-certs\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.915807 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pztq\" (UniqueName: \"kubernetes.io/projected/003334fe-bdcf-4d36-8ba6-705eb6ae4fe3-kube-api-access-5pztq\") pod \"ovn-controller-ovs-xnwzt\" (UID: \"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3\") " pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.924288 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-combined-ca-bundle\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:23 crc kubenswrapper[4708]: I1125 05:53:23.927240 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg7f4\" (UniqueName: \"kubernetes.io/projected/4fe9cd9f-2e87-43fa-b3c1-09560d973e2f-kube-api-access-wg7f4\") pod \"ovn-controller-qc989\" (UID: \"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f\") " pod="openstack/ovn-controller-qc989" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.055377 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.084186 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.607966 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qc989"] Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.622196 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.623649 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.625400 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.625559 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-p2tfp" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.625863 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.625902 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.633995 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.634264 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.648432 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-xnwzt"] Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813126 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813196 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813248 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813290 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813310 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813328 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813352 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n825h\" (UniqueName: \"kubernetes.io/projected/131d2c08-5b83-48df-914b-1f5c28b0ba47-kube-api-access-n825h\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.813373 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-config\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918232 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918435 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918483 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918553 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918634 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n825h\" (UniqueName: \"kubernetes.io/projected/131d2c08-5b83-48df-914b-1f5c28b0ba47-kube-api-access-n825h\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918686 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-config\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918782 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.918885 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.919771 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.919863 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.920361 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-config\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.922037 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/131d2c08-5b83-48df-914b-1f5c28b0ba47-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.926942 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.930874 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.939974 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n825h\" (UniqueName: \"kubernetes.io/projected/131d2c08-5b83-48df-914b-1f5c28b0ba47-kube-api-access-n825h\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.945093 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d2c08-5b83-48df-914b-1f5c28b0ba47-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:24 crc kubenswrapper[4708]: I1125 05:53:24.949839 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"131d2c08-5b83-48df-914b-1f5c28b0ba47\") " pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.251218 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.302328 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-bdsjf"] Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.304620 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.311537 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bdsjf"] Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.313907 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.427797 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovn-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.428544 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovs-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.428605 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.428746 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-combined-ca-bundle\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.428804 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7psr\" (UniqueName: \"kubernetes.io/projected/77b982b3-1e42-4b3a-b3b9-e19f104405ed-kube-api-access-x7psr\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.428831 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b982b3-1e42-4b3a-b3b9-e19f104405ed-config\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.531695 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovn-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.531872 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovs-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.531930 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.531965 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-combined-ca-bundle\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.531988 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7psr\" (UniqueName: \"kubernetes.io/projected/77b982b3-1e42-4b3a-b3b9-e19f104405ed-kube-api-access-x7psr\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.532007 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b982b3-1e42-4b3a-b3b9-e19f104405ed-config\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.532701 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/77b982b3-1e42-4b3a-b3b9-e19f104405ed-config\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.532896 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovn-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.532911 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/77b982b3-1e42-4b3a-b3b9-e19f104405ed-ovs-rundir\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.540676 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.541208 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77b982b3-1e42-4b3a-b3b9-e19f104405ed-combined-ca-bundle\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.546428 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7psr\" (UniqueName: \"kubernetes.io/projected/77b982b3-1e42-4b3a-b3b9-e19f104405ed-kube-api-access-x7psr\") pod \"ovn-controller-metrics-bdsjf\" (UID: \"77b982b3-1e42-4b3a-b3b9-e19f104405ed\") " pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:25 crc kubenswrapper[4708]: I1125 05:53:25.625473 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bdsjf" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.246433 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.248042 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.250071 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.250313 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.250506 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-vcbbr" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.253537 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.268227 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375290 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375359 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375436 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375471 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dc58\" (UniqueName: \"kubernetes.io/projected/4c903b56-d8a2-4522-97d1-4b2333be4929-kube-api-access-6dc58\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375613 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-config\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375688 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375735 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.375770 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476505 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476559 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476587 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476609 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dc58\" (UniqueName: \"kubernetes.io/projected/4c903b56-d8a2-4522-97d1-4b2333be4929-kube-api-access-6dc58\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476653 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-config\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476678 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476705 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.476724 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.477088 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.477372 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.477825 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-config\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.478237 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c903b56-d8a2-4522-97d1-4b2333be4929-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.481285 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.483649 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.485086 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c903b56-d8a2-4522-97d1-4b2333be4929-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.493650 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dc58\" (UniqueName: \"kubernetes.io/projected/4c903b56-d8a2-4522-97d1-4b2333be4929-kube-api-access-6dc58\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.506022 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4c903b56-d8a2-4522-97d1-4b2333be4929\") " pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: W1125 05:53:27.525928 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fe9cd9f_2e87_43fa_b3c1_09560d973e2f.slice/crio-d3224302863243a0a423b1f23389472bd4d2d604f2f701db5aab0e80dc598499 WatchSource:0}: Error finding container d3224302863243a0a423b1f23389472bd4d2d604f2f701db5aab0e80dc598499: Status 404 returned error can't find the container with id d3224302863243a0a423b1f23389472bd4d2d604f2f701db5aab0e80dc598499 Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.571363 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.617613 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qc989" event={"ID":"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f","Type":"ContainerStarted","Data":"d3224302863243a0a423b1f23389472bd4d2d604f2f701db5aab0e80dc598499"} Nov 25 05:53:27 crc kubenswrapper[4708]: I1125 05:53:27.620060 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xnwzt" event={"ID":"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3","Type":"ContainerStarted","Data":"cdefc677639f91ba9abd02eef2261f8a63c53664d66698ba4ee42f2bdaf14d4f"} Nov 25 05:53:30 crc kubenswrapper[4708]: I1125 05:53:30.048457 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.635445 4708 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-mariadb:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.635858 4708 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-mariadb:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.636029 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.rdoproject.org/podified-master-centos9/openstack-mariadb:2cf1dc4bf18c6d57e990b3cd04e8ec78,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tp6sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.637251 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.648409 4708 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.648451 4708 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.648625 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cdw2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-6c965d56d7-dkqw7_openstack(2ef49314-e8b4-4aee-8a73-e10b2dc6e461): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.649894 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" podUID="2ef49314-e8b4-4aee-8a73-e10b2dc6e461" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.657887 4708 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.657948 4708 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.658100 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdn2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bd6d59bf-n6v9f_openstack(3226b19a-018d-47fd-8118-abaa87271f85): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.659602 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" podUID="3226b19a-018d-47fd-8118-abaa87271f85" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.732318 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos9/openstack-mariadb:2cf1dc4bf18c6d57e990b3cd04e8ec78\\\"\"" pod="openstack/openstack-galera-0" podUID="33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.732627 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78\\\"\"" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" podUID="3226b19a-018d-47fd-8118-abaa87271f85" Nov 25 05:53:42 crc kubenswrapper[4708]: E1125 05:53:42.733196 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78\\\"\"" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" podUID="2ef49314-e8b4-4aee-8a73-e10b2dc6e461" Nov 25 05:53:43 crc kubenswrapper[4708]: I1125 05:53:43.038724 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 05:53:43 crc kubenswrapper[4708]: E1125 05:53:43.236801 4708 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-memcached:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:43 crc kubenswrapper[4708]: E1125 05:53:43.237216 4708 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos9/openstack-memcached:2cf1dc4bf18c6d57e990b3cd04e8ec78" Nov 25 05:53:43 crc kubenswrapper[4708]: E1125 05:53:43.237431 4708 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.rdoproject.org/podified-master-centos9/openstack-memcached:2cf1dc4bf18c6d57e990b3cd04e8ec78,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:ncdh655h74hb9h58bh5c7h59h667h598h584h58fh5b6h669h566h676h8dhb9h5cfh549h658h8dh665hbh89h658hcch658h587h5f6hfbh8bh664q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ttz7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 05:53:43 crc kubenswrapper[4708]: E1125 05:53:43.238710 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8" Nov 25 05:53:43 crc kubenswrapper[4708]: E1125 05:53:43.746819 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos9/openstack-memcached:2cf1dc4bf18c6d57e990b3cd04e8ec78\\\"\"" pod="openstack/memcached-0" podUID="c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8" Nov 25 05:53:44 crc kubenswrapper[4708]: W1125 05:53:44.200158 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131d2c08_5b83_48df_914b_1f5c28b0ba47.slice/crio-56969432bad7417b9fa20055f382e5c58f3d94b1a656a5920e0b278f624cf0e5 WatchSource:0}: Error finding container 56969432bad7417b9fa20055f382e5c58f3d94b1a656a5920e0b278f624cf0e5: Status 404 returned error can't find the container with id 56969432bad7417b9fa20055f382e5c58f3d94b1a656a5920e0b278f624cf0e5 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.357817 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.358250 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.358312 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.359016 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.359077 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977" gracePeriod=600 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.579639 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bdsjf"] Nov 25 05:53:44 crc kubenswrapper[4708]: W1125 05:53:44.596352 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77b982b3_1e42_4b3a_b3b9_e19f104405ed.slice/crio-69435b7e6f8c6df79d7ed78793eb2307bfb24658bd77fc99ddee6a030f631442 WatchSource:0}: Error finding container 69435b7e6f8c6df79d7ed78793eb2307bfb24658bd77fc99ddee6a030f631442: Status 404 returned error can't find the container with id 69435b7e6f8c6df79d7ed78793eb2307bfb24658bd77fc99ddee6a030f631442 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.753300 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bdsjf" event={"ID":"77b982b3-1e42-4b3a-b3b9-e19f104405ed","Type":"ContainerStarted","Data":"69435b7e6f8c6df79d7ed78793eb2307bfb24658bd77fc99ddee6a030f631442"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.756190 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"131d2c08-5b83-48df-914b-1f5c28b0ba47","Type":"ContainerStarted","Data":"56969432bad7417b9fa20055f382e5c58f3d94b1a656a5920e0b278f624cf0e5"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.758460 4708 generic.go:334] "Generic (PLEG): container finished" podID="e291e86e-e697-4d41-8e2d-495f975371f0" containerID="b37177be6283fca36b855c38bebe3df2590c97c2794bf24a0c697e9e93ce7f4a" exitCode=0 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.758573 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" event={"ID":"e291e86e-e697-4d41-8e2d-495f975371f0","Type":"ContainerDied","Data":"b37177be6283fca36b855c38bebe3df2590c97c2794bf24a0c697e9e93ce7f4a"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.760177 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b8379446-127e-4f18-b730-085e4b80e60b","Type":"ContainerStarted","Data":"3b255fd96faa50c278f201e3544eeea1de3ae74b3062f7b0d4128b7c1589502a"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.763001 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977" exitCode=0 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.763076 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.763118 4708 scope.go:117] "RemoveContainer" containerID="1d1fef53ae1c52c3e7653dd8644dfa41402a30849b6c08c4c87f0e94c196172f" Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.765381 4708 generic.go:334] "Generic (PLEG): container finished" podID="ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" containerID="f56f7c09c3bd5c49ce31f99f04c2742812ee23c58ede873824ad1fe795136321" exitCode=0 Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.765407 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f48db6df-z84xs" event={"ID":"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39","Type":"ContainerDied","Data":"f56f7c09c3bd5c49ce31f99f04c2742812ee23c58ede873824ad1fe795136321"} Nov 25 05:53:44 crc kubenswrapper[4708]: I1125 05:53:44.907789 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 05:53:44 crc kubenswrapper[4708]: W1125 05:53:44.981342 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c903b56_d8a2_4522_97d1_4b2333be4929.slice/crio-b41c416e592c3f1b279074c47cff2e03babd3a6eaf04e8f31e60268dd09ef2aa WatchSource:0}: Error finding container b41c416e592c3f1b279074c47cff2e03babd3a6eaf04e8f31e60268dd09ef2aa: Status 404 returned error can't find the container with id b41c416e592c3f1b279074c47cff2e03babd3a6eaf04e8f31e60268dd09ef2aa Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.058232 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.160560 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config\") pod \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.160658 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc\") pod \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.160784 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l6dvw\" (UniqueName: \"kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw\") pod \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\" (UID: \"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39\") " Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.179388 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw" (OuterVolumeSpecName: "kube-api-access-l6dvw") pod "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" (UID: "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39"). InnerVolumeSpecName "kube-api-access-l6dvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.194893 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.262715 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l6dvw\" (UniqueName: \"kubernetes.io/projected/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-kube-api-access-l6dvw\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.285556 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config" (OuterVolumeSpecName: "config") pod "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" (UID: "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.292657 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" (UID: "ce3c89d2-03cc-4a71-ae7e-d49a930c9b39"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.364961 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config\") pod \"e291e86e-e697-4d41-8e2d-495f975371f0\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.365022 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df6f8\" (UniqueName: \"kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8\") pod \"e291e86e-e697-4d41-8e2d-495f975371f0\" (UID: \"e291e86e-e697-4d41-8e2d-495f975371f0\") " Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.365865 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.365893 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.369493 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8" (OuterVolumeSpecName: "kube-api-access-df6f8") pod "e291e86e-e697-4d41-8e2d-495f975371f0" (UID: "e291e86e-e697-4d41-8e2d-495f975371f0"). InnerVolumeSpecName "kube-api-access-df6f8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.467379 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df6f8\" (UniqueName: \"kubernetes.io/projected/e291e86e-e697-4d41-8e2d-495f975371f0-kube-api-access-df6f8\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.479285 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config" (OuterVolumeSpecName: "config") pod "e291e86e-e697-4d41-8e2d-495f975371f0" (UID: "e291e86e-e697-4d41-8e2d-495f975371f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.570117 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e291e86e-e697-4d41-8e2d-495f975371f0-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.776746 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qc989" event={"ID":"4fe9cd9f-2e87-43fa-b3c1-09560d973e2f","Type":"ContainerStarted","Data":"323b83b9510f696bf3d26df1155fe7e77fd0c29a51ec1f57d0262ffbe7122a33"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.777104 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-qc989" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.778723 4708 generic.go:334] "Generic (PLEG): container finished" podID="003334fe-bdcf-4d36-8ba6-705eb6ae4fe3" containerID="cb1454905dbc8e62f41f93284d3358e9f6edc68c39821d80a01b7dc6c0824528" exitCode=0 Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.778778 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xnwzt" event={"ID":"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3","Type":"ContainerDied","Data":"cb1454905dbc8e62f41f93284d3358e9f6edc68c39821d80a01b7dc6c0824528"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.782450 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4c903b56-d8a2-4522-97d1-4b2333be4929","Type":"ContainerStarted","Data":"b41c416e592c3f1b279074c47cff2e03babd3a6eaf04e8f31e60268dd09ef2aa"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.798307 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.798291 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-qc989" podStartSLOduration=5.838606488 podStartE2EDuration="22.798269549s" podCreationTimestamp="2025-11-25 05:53:23 +0000 UTC" firstStartedPulling="2025-11-25 05:53:27.529736385 +0000 UTC m=+748.938569771" lastFinishedPulling="2025-11-25 05:53:44.489399447 +0000 UTC m=+765.898232832" observedRunningTime="2025-11-25 05:53:45.789285335 +0000 UTC m=+767.198118721" watchObservedRunningTime="2025-11-25 05:53:45.798269549 +0000 UTC m=+767.207102935" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.802596 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f48db6df-z84xs" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.802667 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f48db6df-z84xs" event={"ID":"ce3c89d2-03cc-4a71-ae7e-d49a930c9b39","Type":"ContainerDied","Data":"bf000bf6560685b48687d24220a4d2b602dde428914039210413a5142f181c5e"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.802744 4708 scope.go:117] "RemoveContainer" containerID="f56f7c09c3bd5c49ce31f99f04c2742812ee23c58ede873824ad1fe795136321" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.812714 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerStarted","Data":"4a020cba17aff994f1bdc0fc2031f1ebde1a2310c8723fd78aeaf80fe4748d3b"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.816747 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" event={"ID":"e291e86e-e697-4d41-8e2d-495f975371f0","Type":"ContainerDied","Data":"2e4d873b23e244cef331a9f18bd12cba509552822fe2b628a2e72070e7e051b2"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.816775 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56fbcfc7-dhtdd" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.821126 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerStarted","Data":"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5"} Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.839880 4708 scope.go:117] "RemoveContainer" containerID="b37177be6283fca36b855c38bebe3df2590c97c2794bf24a0c697e9e93ce7f4a" Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.906584 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.913317 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f48db6df-z84xs"] Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.921821 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:45 crc kubenswrapper[4708]: I1125 05:53:45.925322 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56fbcfc7-dhtdd"] Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.834118 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xnwzt" event={"ID":"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3","Type":"ContainerStarted","Data":"d921b2c62da5c29a38586ce556e00908be242d65b3a7bdde5bd7e3f0d889ffed"} Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.834669 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-xnwzt" event={"ID":"003334fe-bdcf-4d36-8ba6-705eb6ae4fe3","Type":"ContainerStarted","Data":"79c00cae1dee43eceae525855d6ad618343b944bca1e90e2297ddaf54c884a69"} Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.834701 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.840843 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"131d2c08-5b83-48df-914b-1f5c28b0ba47","Type":"ContainerStarted","Data":"2b4865537408b16a0696664e3dff2da58b3a23234fabde6606fc29b302af25a7"} Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.855253 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-xnwzt" podStartSLOduration=6.943524664 podStartE2EDuration="23.85521737s" podCreationTimestamp="2025-11-25 05:53:23 +0000 UTC" firstStartedPulling="2025-11-25 05:53:27.527015329 +0000 UTC m=+748.935848705" lastFinishedPulling="2025-11-25 05:53:44.438708025 +0000 UTC m=+765.847541411" observedRunningTime="2025-11-25 05:53:46.85270153 +0000 UTC m=+768.261534916" watchObservedRunningTime="2025-11-25 05:53:46.85521737 +0000 UTC m=+768.264050755" Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.900965 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" path="/var/lib/kubelet/pods/ce3c89d2-03cc-4a71-ae7e-d49a930c9b39/volumes" Nov 25 05:53:46 crc kubenswrapper[4708]: I1125 05:53:46.901593 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e291e86e-e697-4d41-8e2d-495f975371f0" path="/var/lib/kubelet/pods/e291e86e-e697-4d41-8e2d-495f975371f0/volumes" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.396673 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.426547 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:47 crc kubenswrapper[4708]: E1125 05:53:47.426977 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.426991 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: E1125 05:53:47.427013 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e291e86e-e697-4d41-8e2d-495f975371f0" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.427018 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e291e86e-e697-4d41-8e2d-495f975371f0" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.429723 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce3c89d2-03cc-4a71-ae7e-d49a930c9b39" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.429750 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e291e86e-e697-4d41-8e2d-495f975371f0" containerName="init" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.431074 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.434705 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.435394 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.515832 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg6ww\" (UniqueName: \"kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.515931 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.515986 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.516021 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.619488 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg6ww\" (UniqueName: \"kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.619877 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.620017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.620072 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.621121 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.621269 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.621757 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.643327 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg6ww\" (UniqueName: \"kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww\") pod \"dnsmasq-dns-7586dcdb7c-d7pnd\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.676656 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.687116 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.706210 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.708459 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.720233 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.756157 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.823670 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.823754 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.823833 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz6kn\" (UniqueName: \"kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.823854 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.823917 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.852675 4708 generic.go:334] "Generic (PLEG): container finished" podID="b8379446-127e-4f18-b730-085e4b80e60b" containerID="3b255fd96faa50c278f201e3544eeea1de3ae74b3062f7b0d4128b7c1589502a" exitCode=0 Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.852895 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b8379446-127e-4f18-b730-085e4b80e60b","Type":"ContainerDied","Data":"3b255fd96faa50c278f201e3544eeea1de3ae74b3062f7b0d4128b7c1589502a"} Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.853591 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.925387 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz6kn\" (UniqueName: \"kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.926333 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.926590 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.927173 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.927215 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.927691 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.927777 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.929063 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.929093 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:47 crc kubenswrapper[4708]: I1125 05:53:47.942880 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz6kn\" (UniqueName: \"kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn\") pod \"dnsmasq-dns-748bccdb67-g5b26\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.028439 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.203776 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.228478 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.333698 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config\") pod \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.333785 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config\") pod \"3226b19a-018d-47fd-8118-abaa87271f85\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.333832 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdn2h\" (UniqueName: \"kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h\") pod \"3226b19a-018d-47fd-8118-abaa87271f85\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.333910 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdw2q\" (UniqueName: \"kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q\") pod \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.333957 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc\") pod \"3226b19a-018d-47fd-8118-abaa87271f85\" (UID: \"3226b19a-018d-47fd-8118-abaa87271f85\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.334180 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc\") pod \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\" (UID: \"2ef49314-e8b4-4aee-8a73-e10b2dc6e461\") " Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.335585 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config" (OuterVolumeSpecName: "config") pod "2ef49314-e8b4-4aee-8a73-e10b2dc6e461" (UID: "2ef49314-e8b4-4aee-8a73-e10b2dc6e461"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.335596 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2ef49314-e8b4-4aee-8a73-e10b2dc6e461" (UID: "2ef49314-e8b4-4aee-8a73-e10b2dc6e461"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.336161 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3226b19a-018d-47fd-8118-abaa87271f85" (UID: "3226b19a-018d-47fd-8118-abaa87271f85"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.336403 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config" (OuterVolumeSpecName: "config") pod "3226b19a-018d-47fd-8118-abaa87271f85" (UID: "3226b19a-018d-47fd-8118-abaa87271f85"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.342683 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q" (OuterVolumeSpecName: "kube-api-access-cdw2q") pod "2ef49314-e8b4-4aee-8a73-e10b2dc6e461" (UID: "2ef49314-e8b4-4aee-8a73-e10b2dc6e461"). InnerVolumeSpecName "kube-api-access-cdw2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.348698 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h" (OuterVolumeSpecName: "kube-api-access-sdn2h") pod "3226b19a-018d-47fd-8118-abaa87271f85" (UID: "3226b19a-018d-47fd-8118-abaa87271f85"). InnerVolumeSpecName "kube-api-access-sdn2h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.369975 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:48 crc kubenswrapper[4708]: W1125 05:53:48.405037 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice/crio-83d393e612d68fc4ab7dd823053725b6b067afe29bcdf0cf1d97d6209826a136 WatchSource:0}: Error finding container 83d393e612d68fc4ab7dd823053725b6b067afe29bcdf0cf1d97d6209826a136: Status 404 returned error can't find the container with id 83d393e612d68fc4ab7dd823053725b6b067afe29bcdf0cf1d97d6209826a136 Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.436974 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.437006 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.437017 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdn2h\" (UniqueName: \"kubernetes.io/projected/3226b19a-018d-47fd-8118-abaa87271f85-kube-api-access-sdn2h\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.437029 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdw2q\" (UniqueName: \"kubernetes.io/projected/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-kube-api-access-cdw2q\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.437039 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3226b19a-018d-47fd-8118-abaa87271f85-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.437050 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2ef49314-e8b4-4aee-8a73-e10b2dc6e461-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.446805 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.863426 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerID="1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0" exitCode=0 Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.863544 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" event={"ID":"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d","Type":"ContainerDied","Data":"1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.863923 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" event={"ID":"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d","Type":"ContainerStarted","Data":"953f487d847a9cd34a17081f8ebd4e3fa19da6108f232fa58643cafa29d02852"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.866672 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"131d2c08-5b83-48df-914b-1f5c28b0ba47","Type":"ContainerStarted","Data":"2eb549a60430e26a904a514133960ff4df36fb7fc481cf3e040d25bf9a7730a8"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.869108 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"b8379446-127e-4f18-b730-085e4b80e60b","Type":"ContainerStarted","Data":"a8f31e765cfb3329e8a5d629cb43b3cd0a717c4ed51e391aacacf6b33114710e"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.871266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" event={"ID":"3226b19a-018d-47fd-8118-abaa87271f85","Type":"ContainerDied","Data":"9ce7494b2985534bde9195b8e1f7baeb2a58563f53b3dcf2d5168310c35719f1"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.871373 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bd6d59bf-n6v9f" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.876414 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4c903b56-d8a2-4522-97d1-4b2333be4929","Type":"ContainerStarted","Data":"7cb66a42614bfd6c9d1e32fa74dbf1e50b0806def77147de73f4a8e76b372e0c"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.876449 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4c903b56-d8a2-4522-97d1-4b2333be4929","Type":"ContainerStarted","Data":"e01de224b5586a36cde3165e2d085f98be48f2c190ec73a369d27800555ac65e"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.882345 4708 generic.go:334] "Generic (PLEG): container finished" podID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerID="62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a" exitCode=0 Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.882464 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" event={"ID":"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf","Type":"ContainerDied","Data":"62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.882500 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" event={"ID":"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf","Type":"ContainerStarted","Data":"83d393e612d68fc4ab7dd823053725b6b067afe29bcdf0cf1d97d6209826a136"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.897054 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.913813 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c965d56d7-dkqw7" event={"ID":"2ef49314-e8b4-4aee-8a73-e10b2dc6e461","Type":"ContainerDied","Data":"e26ef4c662c1c8cff78185824b9171d794c9ebcace3433d1461b89876ca9d2cf"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.923145 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bdsjf" event={"ID":"77b982b3-1e42-4b3a-b3b9-e19f104405ed","Type":"ContainerStarted","Data":"bfb7b7629b640eab22d7be990855d77dab6bcd4e20f7a6f321ff82776f2755cd"} Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.931063 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.781401703 podStartE2EDuration="22.931043056s" podCreationTimestamp="2025-11-25 05:53:26 +0000 UTC" firstStartedPulling="2025-11-25 05:53:44.984697992 +0000 UTC m=+766.393531377" lastFinishedPulling="2025-11-25 05:53:48.134339344 +0000 UTC m=+769.543172730" observedRunningTime="2025-11-25 05:53:48.913258615 +0000 UTC m=+770.322092001" watchObservedRunningTime="2025-11-25 05:53:48.931043056 +0000 UTC m=+770.339876442" Nov 25 05:53:48 crc kubenswrapper[4708]: I1125 05:53:48.996463 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=7.087424356 podStartE2EDuration="32.99643558s" podCreationTimestamp="2025-11-25 05:53:16 +0000 UTC" firstStartedPulling="2025-11-25 05:53:18.526768414 +0000 UTC m=+739.935601800" lastFinishedPulling="2025-11-25 05:53:44.435779639 +0000 UTC m=+765.844613024" observedRunningTime="2025-11-25 05:53:48.941273769 +0000 UTC m=+770.350107154" watchObservedRunningTime="2025-11-25 05:53:48.99643558 +0000 UTC m=+770.405268966" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.014769 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=22.062916808 podStartE2EDuration="26.014749189s" podCreationTimestamp="2025-11-25 05:53:23 +0000 UTC" firstStartedPulling="2025-11-25 05:53:44.217870309 +0000 UTC m=+765.626703695" lastFinishedPulling="2025-11-25 05:53:48.16970269 +0000 UTC m=+769.578536076" observedRunningTime="2025-11-25 05:53:48.973271054 +0000 UTC m=+770.382104440" watchObservedRunningTime="2025-11-25 05:53:49.014749189 +0000 UTC m=+770.423582576" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.021540 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-bdsjf" podStartSLOduration=20.455323219 podStartE2EDuration="24.02151286s" podCreationTimestamp="2025-11-25 05:53:25 +0000 UTC" firstStartedPulling="2025-11-25 05:53:44.608233021 +0000 UTC m=+766.017066406" lastFinishedPulling="2025-11-25 05:53:48.174422661 +0000 UTC m=+769.583256047" observedRunningTime="2025-11-25 05:53:48.999851996 +0000 UTC m=+770.408685402" watchObservedRunningTime="2025-11-25 05:53:49.02151286 +0000 UTC m=+770.430346246" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.047354 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.053647 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bd6d59bf-n6v9f"] Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.082205 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.086791 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c965d56d7-dkqw7"] Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.251746 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.288115 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.932738 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" event={"ID":"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf","Type":"ContainerStarted","Data":"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1"} Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.933127 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.935545 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" event={"ID":"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d","Type":"ContainerStarted","Data":"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578"} Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.936025 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.936079 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.954739 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" podStartSLOduration=2.9547129979999998 podStartE2EDuration="2.954712998s" podCreationTimestamp="2025-11-25 05:53:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:53:49.94724967 +0000 UTC m=+771.356083055" watchObservedRunningTime="2025-11-25 05:53:49.954712998 +0000 UTC m=+771.363546385" Nov 25 05:53:49 crc kubenswrapper[4708]: I1125 05:53:49.969408 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" podStartSLOduration=2.9693974 podStartE2EDuration="2.9693974s" podCreationTimestamp="2025-11-25 05:53:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:53:49.964452455 +0000 UTC m=+771.373285841" watchObservedRunningTime="2025-11-25 05:53:49.9693974 +0000 UTC m=+771.378230787" Nov 25 05:53:50 crc kubenswrapper[4708]: I1125 05:53:50.904255 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ef49314-e8b4-4aee-8a73-e10b2dc6e461" path="/var/lib/kubelet/pods/2ef49314-e8b4-4aee-8a73-e10b2dc6e461/volumes" Nov 25 05:53:50 crc kubenswrapper[4708]: I1125 05:53:50.904682 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3226b19a-018d-47fd-8118-abaa87271f85" path="/var/lib/kubelet/pods/3226b19a-018d-47fd-8118-abaa87271f85/volumes" Nov 25 05:53:51 crc kubenswrapper[4708]: I1125 05:53:51.572443 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:51 crc kubenswrapper[4708]: I1125 05:53:51.604134 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:51 crc kubenswrapper[4708]: I1125 05:53:51.952389 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:55 crc kubenswrapper[4708]: I1125 05:53:55.288797 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.605180 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.724648 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.726095 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.730121 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-w25nw" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.730235 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.742887 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.742992 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.752481 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.758756 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.895011 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.895085 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-config\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.896116 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.896228 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95bb2\" (UniqueName: \"kubernetes.io/projected/bd40d3a2-afea-42fe-93b4-72902b3046a6-kube-api-access-95bb2\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.896292 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-scripts\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.896583 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.896696 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.897906 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.998867 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-scripts\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999058 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999133 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999195 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999217 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-config\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999248 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:57 crc kubenswrapper[4708]: I1125 05:53:57.999302 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95bb2\" (UniqueName: \"kubernetes.io/projected/bd40d3a2-afea-42fe-93b4-72902b3046a6-kube-api-access-95bb2\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:57.999901 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-scripts\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.000469 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.000698 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd40d3a2-afea-42fe-93b4-72902b3046a6-config\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.006660 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.006804 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.007384 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/bd40d3a2-afea-42fe-93b4-72902b3046a6-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.009881 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc","Type":"ContainerStarted","Data":"707b517c300d0744a352dedde2bc6d57fcc512a14d61a5d1f890060b1cbf1f1c"} Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.016681 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95bb2\" (UniqueName: \"kubernetes.io/projected/bd40d3a2-afea-42fe-93b4-72902b3046a6-kube-api-access-95bb2\") pod \"ovn-northd-0\" (UID: \"bd40d3a2-afea-42fe-93b4-72902b3046a6\") " pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.030713 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.043461 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.044752 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.044790 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.080062 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.080284 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="dnsmasq-dns" containerID="cri-o://9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1" gracePeriod=10 Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.140866 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.427330 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.460452 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 05:53:58 crc kubenswrapper[4708]: W1125 05:53:58.466599 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd40d3a2_afea_42fe_93b4_72902b3046a6.slice/crio-51f9c9e9971cb7a3916fdd191b39218cc11dbb1b2bdbc94920f618b7a7022e13 WatchSource:0}: Error finding container 51f9c9e9971cb7a3916fdd191b39218cc11dbb1b2bdbc94920f618b7a7022e13: Status 404 returned error can't find the container with id 51f9c9e9971cb7a3916fdd191b39218cc11dbb1b2bdbc94920f618b7a7022e13 Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.520719 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb\") pod \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.520811 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config\") pod \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.520883 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc\") pod \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.520967 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg6ww\" (UniqueName: \"kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww\") pod \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\" (UID: \"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf\") " Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.526917 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww" (OuterVolumeSpecName: "kube-api-access-qg6ww") pod "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" (UID: "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf"). InnerVolumeSpecName "kube-api-access-qg6ww". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.558579 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config" (OuterVolumeSpecName: "config") pod "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" (UID: "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.558598 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" (UID: "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.561932 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" (UID: "01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.623831 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.623856 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.623871 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg6ww\" (UniqueName: \"kubernetes.io/projected/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-kube-api-access-qg6ww\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:58 crc kubenswrapper[4708]: I1125 05:53:58.623881 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.020294 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"bd40d3a2-afea-42fe-93b4-72902b3046a6","Type":"ContainerStarted","Data":"51f9c9e9971cb7a3916fdd191b39218cc11dbb1b2bdbc94920f618b7a7022e13"} Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.023321 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8","Type":"ContainerStarted","Data":"5c3fd90d87c9982f18b8232443f3fa7c5a79e66c613fa4c54a61f0966b2f40cf"} Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.023743 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.028298 4708 generic.go:334] "Generic (PLEG): container finished" podID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerID="9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1" exitCode=0 Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.028396 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" event={"ID":"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf","Type":"ContainerDied","Data":"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1"} Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.028449 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" event={"ID":"01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf","Type":"ContainerDied","Data":"83d393e612d68fc4ab7dd823053725b6b067afe29bcdf0cf1d97d6209826a136"} Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.028499 4708 scope.go:117] "RemoveContainer" containerID="9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.028622 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7586dcdb7c-d7pnd" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.051312 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.122683296 podStartE2EDuration="42.051291318s" podCreationTimestamp="2025-11-25 05:53:17 +0000 UTC" firstStartedPulling="2025-11-25 05:53:18.650563354 +0000 UTC m=+740.059396740" lastFinishedPulling="2025-11-25 05:53:58.579171376 +0000 UTC m=+779.988004762" observedRunningTime="2025-11-25 05:53:59.045172582 +0000 UTC m=+780.454005969" watchObservedRunningTime="2025-11-25 05:53:59.051291318 +0000 UTC m=+780.460124704" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.061627 4708 scope.go:117] "RemoveContainer" containerID="62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.075650 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.080824 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7586dcdb7c-d7pnd"] Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.089788 4708 scope.go:117] "RemoveContainer" containerID="9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1" Nov 25 05:53:59 crc kubenswrapper[4708]: E1125 05:53:59.090293 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1\": container with ID starting with 9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1 not found: ID does not exist" containerID="9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.090358 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1"} err="failed to get container status \"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1\": rpc error: code = NotFound desc = could not find container \"9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1\": container with ID starting with 9bc6043675a31b0322192be562158f7d0a5dfe7a1ebb129dae07dad0ebc00fd1 not found: ID does not exist" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.090396 4708 scope.go:117] "RemoveContainer" containerID="62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a" Nov 25 05:53:59 crc kubenswrapper[4708]: E1125 05:53:59.091422 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a\": container with ID starting with 62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a not found: ID does not exist" containerID="62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.091460 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a"} err="failed to get container status \"62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a\": rpc error: code = NotFound desc = could not find container \"62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a\": container with ID starting with 62be19a2dffa0eb64271bbc602805c59aaa920b2ad21bf98bd070fee74a3396a not found: ID does not exist" Nov 25 05:53:59 crc kubenswrapper[4708]: I1125 05:53:59.107547 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 05:54:00 crc kubenswrapper[4708]: I1125 05:54:00.041884 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"bd40d3a2-afea-42fe-93b4-72902b3046a6","Type":"ContainerStarted","Data":"3d425d2975deb23849ef2f1cb19a7dfc9e92f5bbbb3cc9cd2041b180067301c7"} Nov 25 05:54:00 crc kubenswrapper[4708]: I1125 05:54:00.902987 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" path="/var/lib/kubelet/pods/01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf/volumes" Nov 25 05:54:01 crc kubenswrapper[4708]: I1125 05:54:01.051347 4708 generic.go:334] "Generic (PLEG): container finished" podID="33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc" containerID="707b517c300d0744a352dedde2bc6d57fcc512a14d61a5d1f890060b1cbf1f1c" exitCode=0 Nov 25 05:54:01 crc kubenswrapper[4708]: I1125 05:54:01.051423 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc","Type":"ContainerDied","Data":"707b517c300d0744a352dedde2bc6d57fcc512a14d61a5d1f890060b1cbf1f1c"} Nov 25 05:54:01 crc kubenswrapper[4708]: I1125 05:54:01.055150 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"bd40d3a2-afea-42fe-93b4-72902b3046a6","Type":"ContainerStarted","Data":"c94cbff186a9aab690ba87e729e1f62522a2e49b8106e75bf13cdbd777ac7952"} Nov 25 05:54:01 crc kubenswrapper[4708]: I1125 05:54:01.055375 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 05:54:01 crc kubenswrapper[4708]: I1125 05:54:01.096037 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.729724978 podStartE2EDuration="4.09601684s" podCreationTimestamp="2025-11-25 05:53:57 +0000 UTC" firstStartedPulling="2025-11-25 05:53:58.469500672 +0000 UTC m=+779.878334058" lastFinishedPulling="2025-11-25 05:53:59.835792534 +0000 UTC m=+781.244625920" observedRunningTime="2025-11-25 05:54:01.088815103 +0000 UTC m=+782.497648490" watchObservedRunningTime="2025-11-25 05:54:01.09601684 +0000 UTC m=+782.504850225" Nov 25 05:54:02 crc kubenswrapper[4708]: I1125 05:54:02.066420 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc","Type":"ContainerStarted","Data":"715b66e9a9ff0c47169a3fae57c30f74ab870c5f66e41489f523b94633c56e09"} Nov 25 05:54:02 crc kubenswrapper[4708]: I1125 05:54:02.087688 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371989.767109 podStartE2EDuration="47.087666013s" podCreationTimestamp="2025-11-25 05:53:15 +0000 UTC" firstStartedPulling="2025-11-25 05:53:17.032336595 +0000 UTC m=+738.441169981" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:02.08479274 +0000 UTC m=+783.493626126" watchObservedRunningTime="2025-11-25 05:54:02.087666013 +0000 UTC m=+783.496499399" Nov 25 05:54:03 crc kubenswrapper[4708]: I1125 05:54:03.138043 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 05:54:04 crc kubenswrapper[4708]: E1125 05:54:04.431713 4708 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 192.168.26.34:35938->192.168.26.34:46789: write tcp 192.168.26.34:35938->192.168.26.34:46789: write: broken pipe Nov 25 05:54:06 crc kubenswrapper[4708]: I1125 05:54:06.496130 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 05:54:06 crc kubenswrapper[4708]: I1125 05:54:06.496192 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 05:54:06 crc kubenswrapper[4708]: I1125 05:54:06.563497 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.168929 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 05:54:07 crc kubenswrapper[4708]: E1125 05:54:07.613120 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.779455 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-fb32-account-create-5t5ts"] Nov 25 05:54:07 crc kubenswrapper[4708]: E1125 05:54:07.779858 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="dnsmasq-dns" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.779880 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="dnsmasq-dns" Nov 25 05:54:07 crc kubenswrapper[4708]: E1125 05:54:07.779908 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="init" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.779914 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="init" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.780079 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="01fd2cf9-a2a2-4fb3-ab9e-1d60fea51ecf" containerName="dnsmasq-dns" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.780651 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.783263 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.819462 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fb32-account-create-5t5ts"] Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.833455 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-57c9k"] Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.834818 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.839108 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-57c9k"] Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.877371 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.877454 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnk5h\" (UniqueName: \"kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.979741 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.980189 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.980227 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnk5h\" (UniqueName: \"kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.980293 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ntwq\" (UniqueName: \"kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.980718 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:07 crc kubenswrapper[4708]: I1125 05:54:07.998513 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnk5h\" (UniqueName: \"kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h\") pod \"keystone-fb32-account-create-5t5ts\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.019515 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-sxdxd"] Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.020751 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.030911 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sxdxd"] Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.084983 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6ca0-account-create-x9lqj"] Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.087095 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.090128 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.092329 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvj8j\" (UniqueName: \"kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.092468 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.092618 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.092734 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ntwq\" (UniqueName: \"kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.093708 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.096416 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ca0-account-create-x9lqj"] Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.104734 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.107977 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ntwq\" (UniqueName: \"kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq\") pod \"keystone-db-create-57c9k\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.148192 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.194238 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvj8j\" (UniqueName: \"kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.194547 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p28r7\" (UniqueName: \"kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.194593 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.194622 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.195544 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.215658 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvj8j\" (UniqueName: \"kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j\") pod \"placement-db-create-sxdxd\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.296689 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p28r7\" (UniqueName: \"kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.296753 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.297433 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.315416 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p28r7\" (UniqueName: \"kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7\") pod \"placement-6ca0-account-create-x9lqj\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.342272 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.506393 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.517981 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-fb32-account-create-5t5ts"] Nov 25 05:54:08 crc kubenswrapper[4708]: W1125 05:54:08.524279 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc840b75e_0e64_448f_ba51_051c712afe4d.slice/crio-74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8 WatchSource:0}: Error finding container 74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8: Status 404 returned error can't find the container with id 74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8 Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.583864 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-57c9k"] Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.735056 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-sxdxd"] Nov 25 05:54:08 crc kubenswrapper[4708]: W1125 05:54:08.736925 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2076e535_3cc1_49be_8002_cb063502685c.slice/crio-0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440 WatchSource:0}: Error finding container 0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440: Status 404 returned error can't find the container with id 0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440 Nov 25 05:54:08 crc kubenswrapper[4708]: I1125 05:54:08.904937 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ca0-account-create-x9lqj"] Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.132900 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ca0-account-create-x9lqj" event={"ID":"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28","Type":"ContainerStarted","Data":"3bf296dc6137b503c41f0b1816414fee658b22f92c62a442725387b7fc8a7e56"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.132963 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ca0-account-create-x9lqj" event={"ID":"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28","Type":"ContainerStarted","Data":"efe62c32115250d6eb0defa207037dfd8887c6a14c4e31d2d80cd80d7b4a266a"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.134336 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxdxd" event={"ID":"2076e535-3cc1-49be-8002-cb063502685c","Type":"ContainerStarted","Data":"f3bce6e1deb8f06f80c37c18d8ed37882c3b2c6e7a42e95c8a043f3e0366e64d"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.134372 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxdxd" event={"ID":"2076e535-3cc1-49be-8002-cb063502685c","Type":"ContainerStarted","Data":"0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.137223 4708 generic.go:334] "Generic (PLEG): container finished" podID="ce71c684-746d-4e4e-ad21-2f80340fd52b" containerID="575e8ec457441dfba1ae5262caec6f0b82e73b2fcb797c1aca635e06565edce9" exitCode=0 Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.137293 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-57c9k" event={"ID":"ce71c684-746d-4e4e-ad21-2f80340fd52b","Type":"ContainerDied","Data":"575e8ec457441dfba1ae5262caec6f0b82e73b2fcb797c1aca635e06565edce9"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.137313 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-57c9k" event={"ID":"ce71c684-746d-4e4e-ad21-2f80340fd52b","Type":"ContainerStarted","Data":"5d71447da7256d9d443bccea55ab2c566ff3855be2d67743c434743eafa135e9"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.140043 4708 generic.go:334] "Generic (PLEG): container finished" podID="c840b75e-0e64-448f-ba51-051c712afe4d" containerID="992d08eea81ff9987f9e2d771426ee188c50312709dc0d38f1e07503bfadf29b" exitCode=0 Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.140079 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fb32-account-create-5t5ts" event={"ID":"c840b75e-0e64-448f-ba51-051c712afe4d","Type":"ContainerDied","Data":"992d08eea81ff9987f9e2d771426ee188c50312709dc0d38f1e07503bfadf29b"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.140100 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fb32-account-create-5t5ts" event={"ID":"c840b75e-0e64-448f-ba51-051c712afe4d","Type":"ContainerStarted","Data":"74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8"} Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.155055 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6ca0-account-create-x9lqj" podStartSLOduration=1.155027497 podStartE2EDuration="1.155027497s" podCreationTimestamp="2025-11-25 05:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:09.148037218 +0000 UTC m=+790.556870605" watchObservedRunningTime="2025-11-25 05:54:09.155027497 +0000 UTC m=+790.563860883" Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.219233 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-sxdxd" podStartSLOduration=1.219211505 podStartE2EDuration="1.219211505s" podCreationTimestamp="2025-11-25 05:54:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:09.193725185 +0000 UTC m=+790.602558571" watchObservedRunningTime="2025-11-25 05:54:09.219211505 +0000 UTC m=+790.628044890" Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.976156 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.977412 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:09 crc kubenswrapper[4708]: I1125 05:54:09.995924 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.135431 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vj5xd\" (UniqueName: \"kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.135533 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.135571 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.135641 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.135899 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.152681 4708 generic.go:334] "Generic (PLEG): container finished" podID="8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" containerID="3bf296dc6137b503c41f0b1816414fee658b22f92c62a442725387b7fc8a7e56" exitCode=0 Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.152804 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ca0-account-create-x9lqj" event={"ID":"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28","Type":"ContainerDied","Data":"3bf296dc6137b503c41f0b1816414fee658b22f92c62a442725387b7fc8a7e56"} Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.155631 4708 generic.go:334] "Generic (PLEG): container finished" podID="2076e535-3cc1-49be-8002-cb063502685c" containerID="f3bce6e1deb8f06f80c37c18d8ed37882c3b2c6e7a42e95c8a043f3e0366e64d" exitCode=0 Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.156017 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxdxd" event={"ID":"2076e535-3cc1-49be-8002-cb063502685c","Type":"ContainerDied","Data":"f3bce6e1deb8f06f80c37c18d8ed37882c3b2c6e7a42e95c8a043f3e0366e64d"} Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.237913 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vj5xd\" (UniqueName: \"kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.238257 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.238295 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.238359 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.238423 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.239219 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.239276 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.239298 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.239799 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.271671 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vj5xd\" (UniqueName: \"kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd\") pod \"dnsmasq-dns-5c7f7d8965-gr6ck\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.293955 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.491640 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.515141 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.645924 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts\") pod \"ce71c684-746d-4e4e-ad21-2f80340fd52b\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.646235 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnk5h\" (UniqueName: \"kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h\") pod \"c840b75e-0e64-448f-ba51-051c712afe4d\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.646305 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ntwq\" (UniqueName: \"kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq\") pod \"ce71c684-746d-4e4e-ad21-2f80340fd52b\" (UID: \"ce71c684-746d-4e4e-ad21-2f80340fd52b\") " Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.646470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts\") pod \"c840b75e-0e64-448f-ba51-051c712afe4d\" (UID: \"c840b75e-0e64-448f-ba51-051c712afe4d\") " Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.646913 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ce71c684-746d-4e4e-ad21-2f80340fd52b" (UID: "ce71c684-746d-4e4e-ad21-2f80340fd52b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.647455 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c840b75e-0e64-448f-ba51-051c712afe4d" (UID: "c840b75e-0e64-448f-ba51-051c712afe4d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.647490 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ce71c684-746d-4e4e-ad21-2f80340fd52b-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.650821 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h" (OuterVolumeSpecName: "kube-api-access-mnk5h") pod "c840b75e-0e64-448f-ba51-051c712afe4d" (UID: "c840b75e-0e64-448f-ba51-051c712afe4d"). InnerVolumeSpecName "kube-api-access-mnk5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.650915 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq" (OuterVolumeSpecName: "kube-api-access-8ntwq") pod "ce71c684-746d-4e4e-ad21-2f80340fd52b" (UID: "ce71c684-746d-4e4e-ad21-2f80340fd52b"). InnerVolumeSpecName "kube-api-access-8ntwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.748362 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnk5h\" (UniqueName: \"kubernetes.io/projected/c840b75e-0e64-448f-ba51-051c712afe4d-kube-api-access-mnk5h\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.748389 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ntwq\" (UniqueName: \"kubernetes.io/projected/ce71c684-746d-4e4e-ad21-2f80340fd52b-kube-api-access-8ntwq\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.748404 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c840b75e-0e64-448f-ba51-051c712afe4d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:10 crc kubenswrapper[4708]: I1125 05:54:10.754771 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:10 crc kubenswrapper[4708]: W1125 05:54:10.757251 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04b54d2e_8ab3_4a86_802f_8c4abb2feb78.slice/crio-87b607eaf5d85930677eb6469ae2782d6a49b6df7cc287fd73c9e57604668917 WatchSource:0}: Error finding container 87b607eaf5d85930677eb6469ae2782d6a49b6df7cc287fd73c9e57604668917: Status 404 returned error can't find the container with id 87b607eaf5d85930677eb6469ae2782d6a49b6df7cc287fd73c9e57604668917 Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.168696 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-fb32-account-create-5t5ts" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.168674 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-fb32-account-create-5t5ts" event={"ID":"c840b75e-0e64-448f-ba51-051c712afe4d","Type":"ContainerDied","Data":"74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8"} Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.170286 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74a0a8512f91d92a5aa390ab21a35e8ec5a9f91df4e2b370d0f8202f7169dfe8" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.171242 4708 generic.go:334] "Generic (PLEG): container finished" podID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerID="b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4" exitCode=0 Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.171319 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" event={"ID":"04b54d2e-8ab3-4a86-802f-8c4abb2feb78","Type":"ContainerDied","Data":"b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4"} Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.171356 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" event={"ID":"04b54d2e-8ab3-4a86-802f-8c4abb2feb78","Type":"ContainerStarted","Data":"87b607eaf5d85930677eb6469ae2782d6a49b6df7cc287fd73c9e57604668917"} Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.173647 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-57c9k" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.173852 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-57c9k" event={"ID":"ce71c684-746d-4e4e-ad21-2f80340fd52b","Type":"ContainerDied","Data":"5d71447da7256d9d443bccea55ab2c566ff3855be2d67743c434743eafa135e9"} Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.173909 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d71447da7256d9d443bccea55ab2c566ff3855be2d67743c434743eafa135e9" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.190481 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.191005 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c840b75e-0e64-448f-ba51-051c712afe4d" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.191087 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c840b75e-0e64-448f-ba51-051c712afe4d" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.191156 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce71c684-746d-4e4e-ad21-2f80340fd52b" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.191214 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce71c684-746d-4e4e-ad21-2f80340fd52b" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.191454 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c840b75e-0e64-448f-ba51-051c712afe4d" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.191552 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce71c684-746d-4e4e-ad21-2f80340fd52b" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.197042 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.199109 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.200236 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.200360 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-lk4gb" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.200432 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.210345 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.261996 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.262065 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-lock\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.262098 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vs22j\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-kube-api-access-vs22j\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.262208 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.262259 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-cache\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.364487 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-cache\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.364704 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.364770 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-lock\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.364815 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vs22j\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-kube-api-access-vs22j\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.364976 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.365162 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.365273 4708 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.365293 4708 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.365305 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-lock\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.365361 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift podName:6ee6df83-f664-41b7-9c47-594a0920da51 nodeName:}" failed. No retries permitted until 2025-11-25 05:54:11.865332862 +0000 UTC m=+793.274166248 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift") pod "swift-storage-0" (UID: "6ee6df83-f664-41b7-9c47-594a0920da51") : configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.366398 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee6df83-f664-41b7-9c47-594a0920da51-cache\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.389222 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vs22j\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-kube-api-access-vs22j\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.394456 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.498502 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.503301 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.568300 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvj8j\" (UniqueName: \"kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j\") pod \"2076e535-3cc1-49be-8002-cb063502685c\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.568557 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p28r7\" (UniqueName: \"kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7\") pod \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.568627 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts\") pod \"2076e535-3cc1-49be-8002-cb063502685c\" (UID: \"2076e535-3cc1-49be-8002-cb063502685c\") " Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.568697 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts\") pod \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\" (UID: \"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28\") " Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.569005 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2076e535-3cc1-49be-8002-cb063502685c" (UID: "2076e535-3cc1-49be-8002-cb063502685c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.569165 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" (UID: "8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.569279 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2076e535-3cc1-49be-8002-cb063502685c-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.572389 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j" (OuterVolumeSpecName: "kube-api-access-pvj8j") pod "2076e535-3cc1-49be-8002-cb063502685c" (UID: "2076e535-3cc1-49be-8002-cb063502685c"). InnerVolumeSpecName "kube-api-access-pvj8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.573409 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7" (OuterVolumeSpecName: "kube-api-access-p28r7") pod "8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" (UID: "8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28"). InnerVolumeSpecName "kube-api-access-p28r7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.665152 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-l84pl"] Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.665633 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.665655 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.665696 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2076e535-3cc1-49be-8002-cb063502685c" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.665703 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2076e535-3cc1-49be-8002-cb063502685c" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.665921 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" containerName="mariadb-account-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.665944 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="2076e535-3cc1-49be-8002-cb063502685c" containerName="mariadb-database-create" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.666838 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.672097 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.672129 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvj8j\" (UniqueName: \"kubernetes.io/projected/2076e535-3cc1-49be-8002-cb063502685c-kube-api-access-pvj8j\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.672148 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p28r7\" (UniqueName: \"kubernetes.io/projected/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28-kube-api-access-p28r7\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.672535 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.672981 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.691395 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.721743 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-l84pl"] Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.760568 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-p5f7m"] Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.761787 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.767508 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-p5f7m"] Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774469 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ml27\" (UniqueName: \"kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774502 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774608 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774715 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774758 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774783 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.774799 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.808570 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-l84pl"] Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.809399 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-7ml27 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-l84pl" podUID="40510ad8-a208-4f9e-84c4-deb2b4a6acd3" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.876536 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.876596 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htc9h\" (UniqueName: \"kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.876640 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.876841 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.876972 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877099 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877159 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877188 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877218 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877245 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877277 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877292 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877308 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877408 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ml27\" (UniqueName: \"kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877443 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.877731 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.878180 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.878338 4708 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.878360 4708 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.878373 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: E1125 05:54:11.878416 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift podName:6ee6df83-f664-41b7-9c47-594a0920da51 nodeName:}" failed. No retries permitted until 2025-11-25 05:54:12.878399899 +0000 UTC m=+794.287233285 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift") pod "swift-storage-0" (UID: "6ee6df83-f664-41b7-9c47-594a0920da51") : configmap "swift-ring-files" not found Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.880347 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.880728 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.881780 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.892409 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ml27\" (UniqueName: \"kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27\") pod \"swift-ring-rebalance-l84pl\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979105 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979159 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htc9h\" (UniqueName: \"kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979202 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979296 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979360 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979422 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979468 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.979975 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.980441 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.980611 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.983153 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.983405 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.983614 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:11 crc kubenswrapper[4708]: I1125 05:54:11.994403 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htc9h\" (UniqueName: \"kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h\") pod \"swift-ring-rebalance-p5f7m\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.091473 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.192258 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ca0-account-create-x9lqj" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.194150 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ca0-account-create-x9lqj" event={"ID":"8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28","Type":"ContainerDied","Data":"efe62c32115250d6eb0defa207037dfd8887c6a14c4e31d2d80cd80d7b4a266a"} Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.194396 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efe62c32115250d6eb0defa207037dfd8887c6a14c4e31d2d80cd80d7b4a266a" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.204192 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-sxdxd" event={"ID":"2076e535-3cc1-49be-8002-cb063502685c","Type":"ContainerDied","Data":"0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440"} Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.204223 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-sxdxd" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.204231 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0821d4d25ae64109016141eaa0f407e6d016f4d21267ac24d8bab8a79eb7b440" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.208608 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.210187 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" event={"ID":"04b54d2e-8ab3-4a86-802f-8c4abb2feb78","Type":"ContainerStarted","Data":"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74"} Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.210216 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.218603 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.242355 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" podStartSLOduration=3.242336389 podStartE2EDuration="3.242336389s" podCreationTimestamp="2025-11-25 05:54:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:12.23660293 +0000 UTC m=+793.645436316" watchObservedRunningTime="2025-11-25 05:54:12.242336389 +0000 UTC m=+793.651169775" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292051 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292231 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292351 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ml27\" (UniqueName: \"kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292418 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292468 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292550 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.292575 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift\") pod \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\" (UID: \"40510ad8-a208-4f9e-84c4-deb2b4a6acd3\") " Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.293017 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.293368 4708 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.293693 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts" (OuterVolumeSpecName: "scripts") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.296836 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.297046 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.297074 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.297689 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.299866 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27" (OuterVolumeSpecName: "kube-api-access-7ml27") pod "40510ad8-a208-4f9e-84c4-deb2b4a6acd3" (UID: "40510ad8-a208-4f9e-84c4-deb2b4a6acd3"). InnerVolumeSpecName "kube-api-access-7ml27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394742 4708 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394789 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ml27\" (UniqueName: \"kubernetes.io/projected/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-kube-api-access-7ml27\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394805 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394818 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394831 4708 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.394841 4708 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/40510ad8-a208-4f9e-84c4-deb2b4a6acd3-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.504392 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-p5f7m"] Nov 25 05:54:12 crc kubenswrapper[4708]: I1125 05:54:12.903951 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:12 crc kubenswrapper[4708]: E1125 05:54:12.904225 4708 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 05:54:12 crc kubenswrapper[4708]: E1125 05:54:12.904257 4708 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 05:54:12 crc kubenswrapper[4708]: E1125 05:54:12.904350 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift podName:6ee6df83-f664-41b7-9c47-594a0920da51 nodeName:}" failed. No retries permitted until 2025-11-25 05:54:14.904320953 +0000 UTC m=+796.313154339 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift") pod "swift-storage-0" (UID: "6ee6df83-f664-41b7-9c47-594a0920da51") : configmap "swift-ring-files" not found Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.101018 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.220952 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-p5f7m" event={"ID":"14772fed-d981-4b42-ac9b-24dd6ce02b06","Type":"ContainerStarted","Data":"3840a7a571c3039f6cc2c6360359c4a0daedcf038bb18bb69a14f3b00aae88b9"} Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.229229 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-l84pl" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.269948 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-l84pl"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.276157 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-l84pl"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.300152 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-n4fvw"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.301433 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.309387 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n4fvw"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.412357 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-5c67-account-create-k6fhh"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.413547 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.415004 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.415694 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.415751 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b594r\" (UniqueName: \"kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.418936 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5c67-account-create-k6fhh"] Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.517417 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.517484 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b594r\" (UniqueName: \"kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.517667 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.517738 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cglw\" (UniqueName: \"kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.518265 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.536639 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b594r\" (UniqueName: \"kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r\") pod \"glance-db-create-n4fvw\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.617188 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.619419 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.619498 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cglw\" (UniqueName: \"kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.620486 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.635999 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cglw\" (UniqueName: \"kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw\") pod \"glance-5c67-account-create-k6fhh\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:13 crc kubenswrapper[4708]: I1125 05:54:13.726046 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.029430 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-n4fvw"] Nov 25 05:54:14 crc kubenswrapper[4708]: W1125 05:54:14.033001 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4011a3c5_2e4b_410b_88c6_dfc229b330c2.slice/crio-7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd WatchSource:0}: Error finding container 7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd: Status 404 returned error can't find the container with id 7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.090949 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-qc989" podUID="4fe9cd9f-2e87-43fa-b3c1-09560d973e2f" containerName="ovn-controller" probeResult="failure" output=< Nov 25 05:54:14 crc kubenswrapper[4708]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 05:54:14 crc kubenswrapper[4708]: > Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.129398 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-5c67-account-create-k6fhh"] Nov 25 05:54:14 crc kubenswrapper[4708]: W1125 05:54:14.137134 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cb2b1fd_09b1_476b_a6ad_ae22500b5467.slice/crio-c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496 WatchSource:0}: Error finding container c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496: Status 404 returned error can't find the container with id c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496 Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.227750 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n4fvw" event={"ID":"4011a3c5-2e4b-410b-88c6-dfc229b330c2","Type":"ContainerStarted","Data":"7a0ab90e47fcbdb851207e60b1ad0b1bfe91ee1679d8de0431795195bb4eabc5"} Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.227821 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n4fvw" event={"ID":"4011a3c5-2e4b-410b-88c6-dfc229b330c2","Type":"ContainerStarted","Data":"7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd"} Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.230505 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5c67-account-create-k6fhh" event={"ID":"2cb2b1fd-09b1-476b-a6ad-ae22500b5467","Type":"ContainerStarted","Data":"c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496"} Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.248424 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-n4fvw" podStartSLOduration=1.248398599 podStartE2EDuration="1.248398599s" podCreationTimestamp="2025-11-25 05:54:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:14.241417167 +0000 UTC m=+795.650250553" watchObservedRunningTime="2025-11-25 05:54:14.248398599 +0000 UTC m=+795.657231985" Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.905346 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40510ad8-a208-4f9e-84c4-deb2b4a6acd3" path="/var/lib/kubelet/pods/40510ad8-a208-4f9e-84c4-deb2b4a6acd3/volumes" Nov 25 05:54:14 crc kubenswrapper[4708]: I1125 05:54:14.948006 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:14 crc kubenswrapper[4708]: E1125 05:54:14.948398 4708 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 05:54:14 crc kubenswrapper[4708]: E1125 05:54:14.948428 4708 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 05:54:14 crc kubenswrapper[4708]: E1125 05:54:14.948495 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift podName:6ee6df83-f664-41b7-9c47-594a0920da51 nodeName:}" failed. No retries permitted until 2025-11-25 05:54:18.948473736 +0000 UTC m=+800.357307122 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift") pod "swift-storage-0" (UID: "6ee6df83-f664-41b7-9c47-594a0920da51") : configmap "swift-ring-files" not found Nov 25 05:54:15 crc kubenswrapper[4708]: I1125 05:54:15.249634 4708 generic.go:334] "Generic (PLEG): container finished" podID="2cb2b1fd-09b1-476b-a6ad-ae22500b5467" containerID="7e45c73f9d401a7dfc1dfdcb96333776b6974785571bd02b3e9a28860047f8c1" exitCode=0 Nov 25 05:54:15 crc kubenswrapper[4708]: I1125 05:54:15.249773 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5c67-account-create-k6fhh" event={"ID":"2cb2b1fd-09b1-476b-a6ad-ae22500b5467","Type":"ContainerDied","Data":"7e45c73f9d401a7dfc1dfdcb96333776b6974785571bd02b3e9a28860047f8c1"} Nov 25 05:54:15 crc kubenswrapper[4708]: I1125 05:54:15.252438 4708 generic.go:334] "Generic (PLEG): container finished" podID="4011a3c5-2e4b-410b-88c6-dfc229b330c2" containerID="7a0ab90e47fcbdb851207e60b1ad0b1bfe91ee1679d8de0431795195bb4eabc5" exitCode=0 Nov 25 05:54:15 crc kubenswrapper[4708]: I1125 05:54:15.252496 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n4fvw" event={"ID":"4011a3c5-2e4b-410b-88c6-dfc229b330c2","Type":"ContainerDied","Data":"7a0ab90e47fcbdb851207e60b1ad0b1bfe91ee1679d8de0431795195bb4eabc5"} Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.277239 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-n4fvw" event={"ID":"4011a3c5-2e4b-410b-88c6-dfc229b330c2","Type":"ContainerDied","Data":"7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd"} Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.277620 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b4bbea315651f76fdfe4fbcecb2e1ec64f522eb5930dca5f3128cf09a011acd" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.279716 4708 generic.go:334] "Generic (PLEG): container finished" podID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerID="789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5" exitCode=0 Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.279775 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerDied","Data":"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5"} Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.281710 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-5c67-account-create-k6fhh" event={"ID":"2cb2b1fd-09b1-476b-a6ad-ae22500b5467","Type":"ContainerDied","Data":"c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496"} Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.281767 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5848bbb7c8ab7749a6e4feb711dba3c4970fd6dce522af8972a676f6f99b496" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.284079 4708 generic.go:334] "Generic (PLEG): container finished" podID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerID="4a020cba17aff994f1bdc0fc2031f1ebde1a2310c8723fd78aeaf80fe4748d3b" exitCode=0 Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.284165 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerDied","Data":"4a020cba17aff994f1bdc0fc2031f1ebde1a2310c8723fd78aeaf80fe4748d3b"} Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.386168 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.409981 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.504127 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cglw\" (UniqueName: \"kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw\") pod \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.504190 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b594r\" (UniqueName: \"kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r\") pod \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.504285 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts\") pod \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\" (UID: \"2cb2b1fd-09b1-476b-a6ad-ae22500b5467\") " Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.504346 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts\") pod \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\" (UID: \"4011a3c5-2e4b-410b-88c6-dfc229b330c2\") " Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.504839 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4011a3c5-2e4b-410b-88c6-dfc229b330c2" (UID: "4011a3c5-2e4b-410b-88c6-dfc229b330c2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.505091 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2cb2b1fd-09b1-476b-a6ad-ae22500b5467" (UID: "2cb2b1fd-09b1-476b-a6ad-ae22500b5467"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.505777 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.505869 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4011a3c5-2e4b-410b-88c6-dfc229b330c2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.507748 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r" (OuterVolumeSpecName: "kube-api-access-b594r") pod "4011a3c5-2e4b-410b-88c6-dfc229b330c2" (UID: "4011a3c5-2e4b-410b-88c6-dfc229b330c2"). InnerVolumeSpecName "kube-api-access-b594r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.507795 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw" (OuterVolumeSpecName: "kube-api-access-8cglw") pod "2cb2b1fd-09b1-476b-a6ad-ae22500b5467" (UID: "2cb2b1fd-09b1-476b-a6ad-ae22500b5467"). InnerVolumeSpecName "kube-api-access-8cglw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.607974 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cglw\" (UniqueName: \"kubernetes.io/projected/2cb2b1fd-09b1-476b-a6ad-ae22500b5467-kube-api-access-8cglw\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:17 crc kubenswrapper[4708]: I1125 05:54:17.608016 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b594r\" (UniqueName: \"kubernetes.io/projected/4011a3c5-2e4b-410b-88c6-dfc229b330c2-kube-api-access-b594r\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:17 crc kubenswrapper[4708]: E1125 05:54:17.864604 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.294451 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerStarted","Data":"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417"} Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.295252 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.296590 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerStarted","Data":"d57993727053bb552c82679802b72966177b3586dcb69f9d5c201fd750002854"} Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.296826 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.298068 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-n4fvw" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.298152 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-5c67-account-create-k6fhh" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.298075 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-p5f7m" event={"ID":"14772fed-d981-4b42-ac9b-24dd6ce02b06","Type":"ContainerStarted","Data":"c440ee5b17bb5468962b9d22f86d33cbc2dacd3a9c8eb5c355f7220ead17e7fc"} Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.330769 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.474998498 podStartE2EDuration="1m5.330744757s" podCreationTimestamp="2025-11-25 05:53:13 +0000 UTC" firstStartedPulling="2025-11-25 05:53:15.565990846 +0000 UTC m=+736.974824232" lastFinishedPulling="2025-11-25 05:53:44.421737106 +0000 UTC m=+765.830570491" observedRunningTime="2025-11-25 05:54:18.322221281 +0000 UTC m=+799.731054667" watchObservedRunningTime="2025-11-25 05:54:18.330744757 +0000 UTC m=+799.739578143" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.350205 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.126700445 podStartE2EDuration="1m5.35018575s" podCreationTimestamp="2025-11-25 05:53:13 +0000 UTC" firstStartedPulling="2025-11-25 05:53:15.215002056 +0000 UTC m=+736.623835441" lastFinishedPulling="2025-11-25 05:53:44.43848736 +0000 UTC m=+765.847320746" observedRunningTime="2025-11-25 05:54:18.346272037 +0000 UTC m=+799.755105423" watchObservedRunningTime="2025-11-25 05:54:18.35018575 +0000 UTC m=+799.759019135" Nov 25 05:54:18 crc kubenswrapper[4708]: I1125 05:54:18.383648 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-p5f7m" podStartSLOduration=2.723240665 podStartE2EDuration="7.383628468s" podCreationTimestamp="2025-11-25 05:54:11 +0000 UTC" firstStartedPulling="2025-11-25 05:54:12.513654389 +0000 UTC m=+793.922487775" lastFinishedPulling="2025-11-25 05:54:17.174042192 +0000 UTC m=+798.582875578" observedRunningTime="2025-11-25 05:54:18.369993473 +0000 UTC m=+799.778826859" watchObservedRunningTime="2025-11-25 05:54:18.383628468 +0000 UTC m=+799.792461853" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.030147 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:19 crc kubenswrapper[4708]: E1125 05:54:19.030353 4708 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Nov 25 05:54:19 crc kubenswrapper[4708]: E1125 05:54:19.030613 4708 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Nov 25 05:54:19 crc kubenswrapper[4708]: E1125 05:54:19.030676 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift podName:6ee6df83-f664-41b7-9c47-594a0920da51 nodeName:}" failed. No retries permitted until 2025-11-25 05:54:27.030657889 +0000 UTC m=+808.439491275 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift") pod "swift-storage-0" (UID: "6ee6df83-f664-41b7-9c47-594a0920da51") : configmap "swift-ring-files" not found Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.094930 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-qc989" podUID="4fe9cd9f-2e87-43fa-b3c1-09560d973e2f" containerName="ovn-controller" probeResult="failure" output=< Nov 25 05:54:19 crc kubenswrapper[4708]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 05:54:19 crc kubenswrapper[4708]: > Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.124722 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.125734 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-xnwzt" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.355551 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-qc989-config-mpfsb"] Nov 25 05:54:19 crc kubenswrapper[4708]: E1125 05:54:19.355901 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4011a3c5-2e4b-410b-88c6-dfc229b330c2" containerName="mariadb-database-create" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.355914 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4011a3c5-2e4b-410b-88c6-dfc229b330c2" containerName="mariadb-database-create" Nov 25 05:54:19 crc kubenswrapper[4708]: E1125 05:54:19.355942 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cb2b1fd-09b1-476b-a6ad-ae22500b5467" containerName="mariadb-account-create" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.355948 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cb2b1fd-09b1-476b-a6ad-ae22500b5467" containerName="mariadb-account-create" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.356098 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cb2b1fd-09b1-476b-a6ad-ae22500b5467" containerName="mariadb-account-create" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.356111 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4011a3c5-2e4b-410b-88c6-dfc229b330c2" containerName="mariadb-database-create" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.356677 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.358122 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.366306 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qc989-config-mpfsb"] Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.437865 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.438015 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.438083 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfshb\" (UniqueName: \"kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.438141 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.438184 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.438256 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540379 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540479 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540539 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mfshb\" (UniqueName: \"kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540588 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540611 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540651 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540841 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540848 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.540853 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.541439 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.542858 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.558924 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfshb\" (UniqueName: \"kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb\") pod \"ovn-controller-qc989-config-mpfsb\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:19 crc kubenswrapper[4708]: I1125 05:54:19.672756 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:20 crc kubenswrapper[4708]: W1125 05:54:20.129584 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc0fd759_e8d0_4b6e_b9d5_1d8a813d207e.slice/crio-5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a WatchSource:0}: Error finding container 5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a: Status 404 returned error can't find the container with id 5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a Nov 25 05:54:20 crc kubenswrapper[4708]: I1125 05:54:20.129841 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-qc989-config-mpfsb"] Nov 25 05:54:20 crc kubenswrapper[4708]: I1125 05:54:20.296670 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:20 crc kubenswrapper[4708]: I1125 05:54:20.314818 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qc989-config-mpfsb" event={"ID":"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e","Type":"ContainerStarted","Data":"5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a"} Nov 25 05:54:20 crc kubenswrapper[4708]: I1125 05:54:20.348662 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:54:20 crc kubenswrapper[4708]: I1125 05:54:20.349198 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="dnsmasq-dns" containerID="cri-o://ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578" gracePeriod=10 Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.352956 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.353792 4708 generic.go:334] "Generic (PLEG): container finished" podID="fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" containerID="72592f690d89e2a157c7daf0ef72a2c25c4048de91cd5c62518f2fba7db4e947" exitCode=0 Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.354050 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qc989-config-mpfsb" event={"ID":"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e","Type":"ContainerDied","Data":"72592f690d89e2a157c7daf0ef72a2c25c4048de91cd5c62518f2fba7db4e947"} Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.362360 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerID="ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578" exitCode=0 Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.362414 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" event={"ID":"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d","Type":"ContainerDied","Data":"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578"} Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.362444 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" event={"ID":"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d","Type":"ContainerDied","Data":"953f487d847a9cd34a17081f8ebd4e3fa19da6108f232fa58643cafa29d02852"} Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.362445 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-748bccdb67-g5b26" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.362473 4708 scope.go:117] "RemoveContainer" containerID="ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.386221 4708 scope.go:117] "RemoveContainer" containerID="1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.412060 4708 scope.go:117] "RemoveContainer" containerID="ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578" Nov 25 05:54:21 crc kubenswrapper[4708]: E1125 05:54:21.412425 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578\": container with ID starting with ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578 not found: ID does not exist" containerID="ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.412471 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578"} err="failed to get container status \"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578\": rpc error: code = NotFound desc = could not find container \"ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578\": container with ID starting with ec55e874acab83bea10ad8d3a11f35fd962e856395903b15c9fd45c94939b578 not found: ID does not exist" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.412491 4708 scope.go:117] "RemoveContainer" containerID="1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0" Nov 25 05:54:21 crc kubenswrapper[4708]: E1125 05:54:21.412915 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0\": container with ID starting with 1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0 not found: ID does not exist" containerID="1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.412944 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0"} err="failed to get container status \"1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0\": rpc error: code = NotFound desc = could not find container \"1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0\": container with ID starting with 1af4fe93ac08ecc07341f606c9037be7c2b9f13bd72144d51f6b207b7cbf25e0 not found: ID does not exist" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.482189 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc\") pod \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.482387 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz6kn\" (UniqueName: \"kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn\") pod \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.482448 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config\") pod \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.482508 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb\") pod \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.482587 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb\") pod \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\" (UID: \"c5847bc0-a2e9-43ea-bd92-e0fe55a4917d\") " Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.490242 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn" (OuterVolumeSpecName: "kube-api-access-fz6kn") pod "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" (UID: "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d"). InnerVolumeSpecName "kube-api-access-fz6kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.521084 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config" (OuterVolumeSpecName: "config") pod "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" (UID: "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.521411 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" (UID: "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.526497 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" (UID: "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.527562 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" (UID: "c5847bc0-a2e9-43ea-bd92-e0fe55a4917d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.586108 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz6kn\" (UniqueName: \"kubernetes.io/projected/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-kube-api-access-fz6kn\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.586136 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.586151 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.586160 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.586168 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.697714 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:54:21 crc kubenswrapper[4708]: I1125 05:54:21.707182 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-748bccdb67-g5b26"] Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.651051 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702303 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702359 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702495 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702581 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702624 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfshb\" (UniqueName: \"kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702678 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702695 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702731 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run\") pod \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\" (UID: \"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e\") " Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.702838 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run" (OuterVolumeSpecName: "var-run") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.703120 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.703202 4708 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.703216 4708 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.703227 4708 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.703381 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts" (OuterVolumeSpecName: "scripts") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.706310 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb" (OuterVolumeSpecName: "kube-api-access-mfshb") pod "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" (UID: "fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e"). InnerVolumeSpecName "kube-api-access-mfshb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.804917 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mfshb\" (UniqueName: \"kubernetes.io/projected/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-kube-api-access-mfshb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.804951 4708 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.804966 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:22 crc kubenswrapper[4708]: I1125 05:54:22.906731 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" path="/var/lib/kubelet/pods/c5847bc0-a2e9-43ea-bd92-e0fe55a4917d/volumes" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.386740 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-qc989-config-mpfsb" event={"ID":"fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e","Type":"ContainerDied","Data":"5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a"} Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.387152 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5281fbce43c22e9c7c5627e2972a8fdb628eef0e194b0c9429588f91f37d052a" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.386769 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-qc989-config-mpfsb" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.388209 4708 generic.go:334] "Generic (PLEG): container finished" podID="14772fed-d981-4b42-ac9b-24dd6ce02b06" containerID="c440ee5b17bb5468962b9d22f86d33cbc2dacd3a9c8eb5c355f7220ead17e7fc" exitCode=0 Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.388373 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-p5f7m" event={"ID":"14772fed-d981-4b42-ac9b-24dd6ce02b06","Type":"ContainerDied","Data":"c440ee5b17bb5468962b9d22f86d33cbc2dacd3a9c8eb5c355f7220ead17e7fc"} Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.626406 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-qv9pr"] Nov 25 05:54:23 crc kubenswrapper[4708]: E1125 05:54:23.626785 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="dnsmasq-dns" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.626805 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="dnsmasq-dns" Nov 25 05:54:23 crc kubenswrapper[4708]: E1125 05:54:23.626822 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" containerName="ovn-config" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.626829 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" containerName="ovn-config" Nov 25 05:54:23 crc kubenswrapper[4708]: E1125 05:54:23.626851 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="init" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.626859 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="init" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.627031 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" containerName="ovn-config" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.627049 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5847bc0-a2e9-43ea-bd92-e0fe55a4917d" containerName="dnsmasq-dns" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.627647 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.629129 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.629261 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-j2dg4" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.637747 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qv9pr"] Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.722988 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.723084 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxcct\" (UniqueName: \"kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.723117 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.723139 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.724375 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-qc989-config-mpfsb"] Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.732336 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-qc989-config-mpfsb"] Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.825013 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.825100 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxcct\" (UniqueName: \"kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.825132 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.825155 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.830137 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.830266 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.830996 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.857057 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxcct\" (UniqueName: \"kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct\") pod \"glance-db-sync-qv9pr\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:23 crc kubenswrapper[4708]: I1125 05:54:23.942480 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.103687 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-qc989" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.474726 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-qv9pr"] Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.638414 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.748686 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htc9h\" (UniqueName: \"kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.748792 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.748844 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.748926 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.748969 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.749007 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.749099 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf\") pod \"14772fed-d981-4b42-ac9b-24dd6ce02b06\" (UID: \"14772fed-d981-4b42-ac9b-24dd6ce02b06\") " Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.749470 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.749770 4708 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.749759 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.754661 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h" (OuterVolumeSpecName: "kube-api-access-htc9h") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "kube-api-access-htc9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.757305 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.767535 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts" (OuterVolumeSpecName: "scripts") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.770385 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.771822 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "14772fed-d981-4b42-ac9b-24dd6ce02b06" (UID: "14772fed-d981-4b42-ac9b-24dd6ce02b06"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851871 4708 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851906 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htc9h\" (UniqueName: \"kubernetes.io/projected/14772fed-d981-4b42-ac9b-24dd6ce02b06-kube-api-access-htc9h\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851919 4708 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/14772fed-d981-4b42-ac9b-24dd6ce02b06-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851931 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/14772fed-d981-4b42-ac9b-24dd6ce02b06-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851940 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.851949 4708 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/14772fed-d981-4b42-ac9b-24dd6ce02b06-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:24 crc kubenswrapper[4708]: I1125 05:54:24.902562 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e" path="/var/lib/kubelet/pods/fc0fd759-e8d0-4b6e-b9d5-1d8a813d207e/volumes" Nov 25 05:54:25 crc kubenswrapper[4708]: I1125 05:54:25.407199 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-p5f7m" event={"ID":"14772fed-d981-4b42-ac9b-24dd6ce02b06","Type":"ContainerDied","Data":"3840a7a571c3039f6cc2c6360359c4a0daedcf038bb18bb69a14f3b00aae88b9"} Nov 25 05:54:25 crc kubenswrapper[4708]: I1125 05:54:25.407509 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3840a7a571c3039f6cc2c6360359c4a0daedcf038bb18bb69a14f3b00aae88b9" Nov 25 05:54:25 crc kubenswrapper[4708]: I1125 05:54:25.407230 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-p5f7m" Nov 25 05:54:25 crc kubenswrapper[4708]: I1125 05:54:25.410113 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qv9pr" event={"ID":"cc02d86b-b016-4e10-960e-f7237575e10d","Type":"ContainerStarted","Data":"f6598c32f9a45cf357ecd9e4951db7b9cab267a391cf0c3509f717404651c8c0"} Nov 25 05:54:27 crc kubenswrapper[4708]: I1125 05:54:27.092498 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:27 crc kubenswrapper[4708]: I1125 05:54:27.099573 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee6df83-f664-41b7-9c47-594a0920da51-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee6df83-f664-41b7-9c47-594a0920da51\") " pod="openstack/swift-storage-0" Nov 25 05:54:27 crc kubenswrapper[4708]: I1125 05:54:27.120506 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Nov 25 05:54:27 crc kubenswrapper[4708]: I1125 05:54:27.632440 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Nov 25 05:54:28 crc kubenswrapper[4708]: E1125 05:54:28.065646 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:28 crc kubenswrapper[4708]: I1125 05:54:28.443231 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"2425f82a51af42dbbac3b17ac0dba9308fe377bdc58440eee4f8864a5a66b742"} Nov 25 05:54:29 crc kubenswrapper[4708]: I1125 05:54:29.454658 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"9705339e86ee56dc52d3d47e112d296e65c5e6d710efdb00361037ed336a069c"} Nov 25 05:54:29 crc kubenswrapper[4708]: I1125 05:54:29.454984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"948b162d595d0af5734755e670a3fb5ec28a96bbf737596b1d827fe0335e4fca"} Nov 25 05:54:30 crc kubenswrapper[4708]: I1125 05:54:30.464208 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"0f85d60c93576f2345aece9dc5d4db81f8605ce45ca47316354e966c061bc22b"} Nov 25 05:54:30 crc kubenswrapper[4708]: I1125 05:54:30.465039 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"8b4c80618c6b6a6ebf41fdef8f88cd014ced994c3dcdf2208c1c07b013efd5f4"} Nov 25 05:54:31 crc kubenswrapper[4708]: I1125 05:54:31.495338 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"c1e58bb89abf39ed2b29c2dc4bd65c95b381baa40a73a0918faf430376be5a2b"} Nov 25 05:54:31 crc kubenswrapper[4708]: I1125 05:54:31.495711 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"6584f01ad7268ff6a0ef6c2e22ea44e9d2aca739b0dcea0cb7eaceda2da813d9"} Nov 25 05:54:31 crc kubenswrapper[4708]: I1125 05:54:31.495724 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"8158f8c9fee7d19e2fb41c80937f0c4dd30e1979d333f20f7132c5e6c95f6b74"} Nov 25 05:54:32 crc kubenswrapper[4708]: I1125 05:54:32.509923 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"579ed530994a4c225e4f1ee737d72d08ed21814c6ecb4bc188b0acb5a13ce271"} Nov 25 05:54:33 crc kubenswrapper[4708]: I1125 05:54:33.523797 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"cfcb8c7aba1177560fb0bf6a38ac942b87479841bab72c845ede6d85f9d13180"} Nov 25 05:54:33 crc kubenswrapper[4708]: I1125 05:54:33.524107 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"501a6a4187b2f26b39c7287f63477a5f7ffefbd8f9924a85d090462fd0c50387"} Nov 25 05:54:33 crc kubenswrapper[4708]: I1125 05:54:33.524121 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"8da2c92ea9f773836baf42b2aa30ac5b7403ca90bae463968c2ffe25ea4b94bb"} Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.536830 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"8d9050a8a026c14d1ec42bf4d73b3226519f159a06b586624a6d457f9cd0e3a4"} Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.537128 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"ea7dc54b703c0cff817beb4cfb45ab66dbd2bfd8c3cca0757ca4c9d2b421e247"} Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.537139 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"a72dc47dbbd2fb57af178f006a075d5e0782089f1ac0fcadaefd9e42844d4d17"} Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.537148 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee6df83-f664-41b7-9c47-594a0920da51","Type":"ContainerStarted","Data":"4def34edc25d6cecce94a35a26839eff9ed237ab44679fb57613fac86aac6f51"} Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.567958 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=19.243539151 podStartE2EDuration="24.567939819s" podCreationTimestamp="2025-11-25 05:54:10 +0000 UTC" firstStartedPulling="2025-11-25 05:54:27.644446487 +0000 UTC m=+809.053279874" lastFinishedPulling="2025-11-25 05:54:32.968847155 +0000 UTC m=+814.377680542" observedRunningTime="2025-11-25 05:54:34.562898452 +0000 UTC m=+815.971731838" watchObservedRunningTime="2025-11-25 05:54:34.567939819 +0000 UTC m=+815.976773205" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.781948 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:54:34 crc kubenswrapper[4708]: E1125 05:54:34.782325 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14772fed-d981-4b42-ac9b-24dd6ce02b06" containerName="swift-ring-rebalance" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.782343 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="14772fed-d981-4b42-ac9b-24dd6ce02b06" containerName="swift-ring-rebalance" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.782497 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="14772fed-d981-4b42-ac9b-24dd6ce02b06" containerName="swift-ring-rebalance" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.783378 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.786337 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.787746 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.802302 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819665 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819742 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819774 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819879 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819904 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.819960 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g666h\" (UniqueName: \"kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.921616 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.921940 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.922006 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923079 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923125 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923258 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923284 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923341 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g666h\" (UniqueName: \"kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923874 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.923939 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.924625 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:34 crc kubenswrapper[4708]: I1125 05:54:34.940739 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g666h\" (UniqueName: \"kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h\") pod \"dnsmasq-dns-648845bbf5-g7xjf\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:35 crc kubenswrapper[4708]: I1125 05:54:35.101473 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:35 crc kubenswrapper[4708]: I1125 05:54:35.105232 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.407786 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-vw8x4"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.409120 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.417820 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vw8x4"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.450923 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.451172 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5k5l\" (UniqueName: \"kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.507352 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-5hcbt"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.508532 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.518191 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-6f71-account-create-65cgg"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.519466 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.525203 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.527390 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5hcbt"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.538575 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6f71-account-create-65cgg"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552369 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552578 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5k5l\" (UniqueName: \"kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552704 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkltt\" (UniqueName: \"kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552793 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552875 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.552973 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbc5n\" (UniqueName: \"kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.553667 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.572194 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5k5l\" (UniqueName: \"kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l\") pod \"cinder-db-create-vw8x4\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.620718 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-a6c7-account-create-z2ldb"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.624046 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.626970 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.637586 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a6c7-account-create-z2ldb"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654070 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkltt\" (UniqueName: \"kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654121 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654154 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654185 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b7nx\" (UniqueName: \"kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654211 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbc5n\" (UniqueName: \"kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654259 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.654936 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.655145 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.670414 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkltt\" (UniqueName: \"kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt\") pod \"cinder-6f71-account-create-65cgg\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.671274 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbc5n\" (UniqueName: \"kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n\") pod \"barbican-db-create-5hcbt\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.723709 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-b25rc"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.725032 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.727063 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.732310 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b25rc"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.756075 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9hz8\" (UniqueName: \"kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.756206 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b7nx\" (UniqueName: \"kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.756331 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.756411 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.756910 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.772647 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b7nx\" (UniqueName: \"kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx\") pod \"barbican-a6c7-account-create-z2ldb\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.793711 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wxksw"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.794834 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.799122 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.799376 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.802740 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.803030 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sdz4c" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.813344 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wxksw"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.826051 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f6ba-account-create-2c6rf"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.826109 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.826940 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.829866 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.832047 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.837258 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f6ba-account-create-2c6rf"] Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857060 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857096 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmldk\" (UniqueName: \"kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857147 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857214 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9hz8\" (UniqueName: \"kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857255 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857305 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.857355 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4dbl\" (UniqueName: \"kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.858816 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.870068 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9hz8\" (UniqueName: \"kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8\") pod \"neutron-db-create-b25rc\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.948702 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.959159 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.959204 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmldk\" (UniqueName: \"kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.959345 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.959411 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.959461 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4dbl\" (UniqueName: \"kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.962034 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.964812 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.965473 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.978437 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4dbl\" (UniqueName: \"kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl\") pod \"keystone-db-sync-wxksw\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:36 crc kubenswrapper[4708]: I1125 05:54:36.980194 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmldk\" (UniqueName: \"kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk\") pod \"neutron-f6ba-account-create-2c6rf\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:37 crc kubenswrapper[4708]: I1125 05:54:37.041562 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:37 crc kubenswrapper[4708]: I1125 05:54:37.114367 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:37 crc kubenswrapper[4708]: I1125 05:54:37.144156 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:38 crc kubenswrapper[4708]: E1125 05:54:38.277133 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.089540 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wxksw"] Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.101819 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31f8dae9_be16_4ee1_a485_73148c832c47.slice/crio-d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b WatchSource:0}: Error finding container d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b: Status 404 returned error can't find the container with id d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.264154 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-5hcbt"] Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.269925 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda40cd066_4de2_427b_b1f6_207271f43d26.slice/crio-2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93 WatchSource:0}: Error finding container 2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93: Status 404 returned error can't find the container with id 2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93 Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.272105 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08c33523_d685_4072_9673_9989ed7ad578.slice/crio-adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6 WatchSource:0}: Error finding container adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6: Status 404 returned error can't find the container with id adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6 Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.272615 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-a6c7-account-create-z2ldb"] Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.279414 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f6ba-account-create-2c6rf"] Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.400663 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-vw8x4"] Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.409634 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-6f71-account-create-65cgg"] Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.417158 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b25rc"] Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.417953 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c8aaf81_e916_4011_be02_a76c72a956f8.slice/crio-018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526 WatchSource:0}: Error finding container 018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526: Status 404 returned error can't find the container with id 018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526 Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.420509 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod59c87445_1528_4cab_9a2a_8d736c1a6c71.slice/crio-85f7957b059b18a72ee551c44828ddcb26977b9335ab5d89274c91a699ecd802 WatchSource:0}: Error finding container 85f7957b059b18a72ee551c44828ddcb26977b9335ab5d89274c91a699ecd802: Status 404 returned error can't find the container with id 85f7957b059b18a72ee551c44828ddcb26977b9335ab5d89274c91a699ecd802 Nov 25 05:54:42 crc kubenswrapper[4708]: W1125 05:54:42.423648 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c883a5b_0f2a_48a0_b29c_9bfc01fb439d.slice/crio-8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b WatchSource:0}: Error finding container 8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b: Status 404 returned error can't find the container with id 8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.424409 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.625319 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qv9pr" event={"ID":"cc02d86b-b016-4e10-960e-f7237575e10d","Type":"ContainerStarted","Data":"2224ff2439a37fecc13859d4d06accb76f8390714d2ebd25431b1a9ceb639735"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.627395 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" event={"ID":"59c87445-1528-4cab-9a2a-8d736c1a6c71","Type":"ContainerStarted","Data":"85f7957b059b18a72ee551c44828ddcb26977b9335ab5d89274c91a699ecd802"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.629488 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f71-account-create-65cgg" event={"ID":"9c8aaf81-e916-4011-be02-a76c72a956f8","Type":"ContainerStarted","Data":"2e714e36b81afda9a8e91f3780d95e2c3137f3c83f2834e7304dff6f43a348f7"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.629603 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f71-account-create-65cgg" event={"ID":"9c8aaf81-e916-4011-be02-a76c72a956f8","Type":"ContainerStarted","Data":"018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.632130 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6c7-account-create-z2ldb" event={"ID":"a40cd066-4de2-427b-b1f6-207271f43d26","Type":"ContainerStarted","Data":"79113f94b45daa9f043c856c87c0a8bae146cf04c735e7d4eba70a57ff57bd89"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.632156 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6c7-account-create-z2ldb" event={"ID":"a40cd066-4de2-427b-b1f6-207271f43d26","Type":"ContainerStarted","Data":"2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.640761 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b25rc" event={"ID":"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d","Type":"ContainerStarted","Data":"eefe49bf32dbbf9245728e7fe4d324297b299f1dfb6537d8688c8f8fabee86d6"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.640811 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b25rc" event={"ID":"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d","Type":"ContainerStarted","Data":"8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.642066 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-qv9pr" podStartSLOduration=2.377936106 podStartE2EDuration="19.64204933s" podCreationTimestamp="2025-11-25 05:54:23 +0000 UTC" firstStartedPulling="2025-11-25 05:54:24.497207772 +0000 UTC m=+805.906041157" lastFinishedPulling="2025-11-25 05:54:41.761320995 +0000 UTC m=+823.170154381" observedRunningTime="2025-11-25 05:54:42.637538382 +0000 UTC m=+824.046371759" watchObservedRunningTime="2025-11-25 05:54:42.64204933 +0000 UTC m=+824.050882716" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.645952 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vw8x4" event={"ID":"79753d96-259b-4fd5-9dfe-cb27f6ce82d6","Type":"ContainerStarted","Data":"476d5209b4027ed7175a8cd5fc91f95fad8a6285d08e6ca682c529e872d8e508"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.646018 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vw8x4" event={"ID":"79753d96-259b-4fd5-9dfe-cb27f6ce82d6","Type":"ContainerStarted","Data":"91a17f85fc5c64384c16beb71e37991af4ac850eff66b4943b0cc56fe7f9254c"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.648459 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6ba-account-create-2c6rf" event={"ID":"23de5970-5eb9-4f95-b155-e7bc2fdbed7a","Type":"ContainerStarted","Data":"0a59cfb5eff3cd4addb0fa6cbd7300e17102ee45f5f7324b63f61f6d07bc2b05"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.648496 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6ba-account-create-2c6rf" event={"ID":"23de5970-5eb9-4f95-b155-e7bc2fdbed7a","Type":"ContainerStarted","Data":"d3d8664a19a5372acf79e24dabb8f691e783d348ebec05d21c408685ea8a5d18"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.650224 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-6f71-account-create-65cgg" podStartSLOduration=6.650201466 podStartE2EDuration="6.650201466s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:42.64895563 +0000 UTC m=+824.057789016" watchObservedRunningTime="2025-11-25 05:54:42.650201466 +0000 UTC m=+824.059034852" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.657330 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5hcbt" event={"ID":"08c33523-d685-4072-9673-9989ed7ad578","Type":"ContainerStarted","Data":"0ac1c5cf865fd6a4fe3a584bbf1264f52ab177988d9311eb3e68a66a79205eeb"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.657363 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5hcbt" event={"ID":"08c33523-d685-4072-9673-9989ed7ad578","Type":"ContainerStarted","Data":"adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.658666 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wxksw" event={"ID":"31f8dae9-be16-4ee1-a485-73148c832c47","Type":"ContainerStarted","Data":"d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b"} Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.683825 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-a6c7-account-create-z2ldb" podStartSLOduration=6.683809345 podStartE2EDuration="6.683809345s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:42.67609333 +0000 UTC m=+824.084926716" watchObservedRunningTime="2025-11-25 05:54:42.683809345 +0000 UTC m=+824.092642732" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.707453 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-vw8x4" podStartSLOduration=6.707436774 podStartE2EDuration="6.707436774s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:42.700749687 +0000 UTC m=+824.109583073" watchObservedRunningTime="2025-11-25 05:54:42.707436774 +0000 UTC m=+824.116270160" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.718577 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-f6ba-account-create-2c6rf" podStartSLOduration=6.718552945 podStartE2EDuration="6.718552945s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:42.714102391 +0000 UTC m=+824.122935776" watchObservedRunningTime="2025-11-25 05:54:42.718552945 +0000 UTC m=+824.127386331" Nov 25 05:54:42 crc kubenswrapper[4708]: I1125 05:54:42.743774 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-create-b25rc" podStartSLOduration=6.743742976 podStartE2EDuration="6.743742976s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:42.741455117 +0000 UTC m=+824.150288502" watchObservedRunningTime="2025-11-25 05:54:42.743742976 +0000 UTC m=+824.152576352" Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.672951 4708 generic.go:334] "Generic (PLEG): container finished" podID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerID="e2f10046cbe7b3f4ca4d265a8d449c6d7eb8b25a7f96f3ed3c54f6078be495b9" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.673081 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" event={"ID":"59c87445-1528-4cab-9a2a-8d736c1a6c71","Type":"ContainerDied","Data":"e2f10046cbe7b3f4ca4d265a8d449c6d7eb8b25a7f96f3ed3c54f6078be495b9"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.677990 4708 generic.go:334] "Generic (PLEG): container finished" podID="9c8aaf81-e916-4011-be02-a76c72a956f8" containerID="2e714e36b81afda9a8e91f3780d95e2c3137f3c83f2834e7304dff6f43a348f7" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.678080 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f71-account-create-65cgg" event={"ID":"9c8aaf81-e916-4011-be02-a76c72a956f8","Type":"ContainerDied","Data":"2e714e36b81afda9a8e91f3780d95e2c3137f3c83f2834e7304dff6f43a348f7"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.680001 4708 generic.go:334] "Generic (PLEG): container finished" podID="a40cd066-4de2-427b-b1f6-207271f43d26" containerID="79113f94b45daa9f043c856c87c0a8bae146cf04c735e7d4eba70a57ff57bd89" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.680036 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6c7-account-create-z2ldb" event={"ID":"a40cd066-4de2-427b-b1f6-207271f43d26","Type":"ContainerDied","Data":"79113f94b45daa9f043c856c87c0a8bae146cf04c735e7d4eba70a57ff57bd89"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.681545 4708 generic.go:334] "Generic (PLEG): container finished" podID="7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" containerID="eefe49bf32dbbf9245728e7fe4d324297b299f1dfb6537d8688c8f8fabee86d6" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.681604 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b25rc" event={"ID":"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d","Type":"ContainerDied","Data":"eefe49bf32dbbf9245728e7fe4d324297b299f1dfb6537d8688c8f8fabee86d6"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.683129 4708 generic.go:334] "Generic (PLEG): container finished" podID="79753d96-259b-4fd5-9dfe-cb27f6ce82d6" containerID="476d5209b4027ed7175a8cd5fc91f95fad8a6285d08e6ca682c529e872d8e508" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.683206 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vw8x4" event={"ID":"79753d96-259b-4fd5-9dfe-cb27f6ce82d6","Type":"ContainerDied","Data":"476d5209b4027ed7175a8cd5fc91f95fad8a6285d08e6ca682c529e872d8e508"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.684860 4708 generic.go:334] "Generic (PLEG): container finished" podID="08c33523-d685-4072-9673-9989ed7ad578" containerID="0ac1c5cf865fd6a4fe3a584bbf1264f52ab177988d9311eb3e68a66a79205eeb" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.685200 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5hcbt" event={"ID":"08c33523-d685-4072-9673-9989ed7ad578","Type":"ContainerDied","Data":"0ac1c5cf865fd6a4fe3a584bbf1264f52ab177988d9311eb3e68a66a79205eeb"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.688399 4708 generic.go:334] "Generic (PLEG): container finished" podID="23de5970-5eb9-4f95-b155-e7bc2fdbed7a" containerID="0a59cfb5eff3cd4addb0fa6cbd7300e17102ee45f5f7324b63f61f6d07bc2b05" exitCode=0 Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.688457 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6ba-account-create-2c6rf" event={"ID":"23de5970-5eb9-4f95-b155-e7bc2fdbed7a","Type":"ContainerDied","Data":"0a59cfb5eff3cd4addb0fa6cbd7300e17102ee45f5f7324b63f61f6d07bc2b05"} Nov 25 05:54:43 crc kubenswrapper[4708]: I1125 05:54:43.978989 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.115577 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbc5n\" (UniqueName: \"kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n\") pod \"08c33523-d685-4072-9673-9989ed7ad578\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.116403 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts\") pod \"08c33523-d685-4072-9673-9989ed7ad578\" (UID: \"08c33523-d685-4072-9673-9989ed7ad578\") " Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.117116 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "08c33523-d685-4072-9673-9989ed7ad578" (UID: "08c33523-d685-4072-9673-9989ed7ad578"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.117466 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/08c33523-d685-4072-9673-9989ed7ad578-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.125059 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n" (OuterVolumeSpecName: "kube-api-access-xbc5n") pod "08c33523-d685-4072-9673-9989ed7ad578" (UID: "08c33523-d685-4072-9673-9989ed7ad578"). InnerVolumeSpecName "kube-api-access-xbc5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.219910 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbc5n\" (UniqueName: \"kubernetes.io/projected/08c33523-d685-4072-9673-9989ed7ad578-kube-api-access-xbc5n\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.701455 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-5hcbt" event={"ID":"08c33523-d685-4072-9673-9989ed7ad578","Type":"ContainerDied","Data":"adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6"} Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.701503 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-5hcbt" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.701540 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adbc833ef0d14ac71fcff418e8c49bf5c32e6f4cddedac44d29c0d5479b250d6" Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.704709 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" event={"ID":"59c87445-1528-4cab-9a2a-8d736c1a6c71","Type":"ContainerStarted","Data":"f38ca8faa43e41d183bca58a0da7dd333b2abd599cdad0cc03e37b376b19542d"} Nov 25 05:54:44 crc kubenswrapper[4708]: I1125 05:54:44.725407 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" podStartSLOduration=10.725386326 podStartE2EDuration="10.725386326s" podCreationTimestamp="2025-11-25 05:54:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:44.723715858 +0000 UTC m=+826.132549245" watchObservedRunningTime="2025-11-25 05:54:44.725386326 +0000 UTC m=+826.134219712" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.045905 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.104556 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.136827 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts\") pod \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.137015 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmldk\" (UniqueName: \"kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk\") pod \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\" (UID: \"23de5970-5eb9-4f95-b155-e7bc2fdbed7a\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.137627 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "23de5970-5eb9-4f95-b155-e7bc2fdbed7a" (UID: "23de5970-5eb9-4f95-b155-e7bc2fdbed7a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.142343 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk" (OuterVolumeSpecName: "kube-api-access-pmldk") pod "23de5970-5eb9-4f95-b155-e7bc2fdbed7a" (UID: "23de5970-5eb9-4f95-b155-e7bc2fdbed7a"). InnerVolumeSpecName "kube-api-access-pmldk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.196449 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.209102 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.215515 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.234537 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.241140 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.241181 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmldk\" (UniqueName: \"kubernetes.io/projected/23de5970-5eb9-4f95-b155-e7bc2fdbed7a-kube-api-access-pmldk\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.342814 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5k5l\" (UniqueName: \"kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l\") pod \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.342862 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts\") pod \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.342901 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts\") pod \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\" (UID: \"79753d96-259b-4fd5-9dfe-cb27f6ce82d6\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.343576 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" (UID: "7c883a5b-0f2a-48a0-b29c-9bfc01fb439d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.343614 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "79753d96-259b-4fd5-9dfe-cb27f6ce82d6" (UID: "79753d96-259b-4fd5-9dfe-cb27f6ce82d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.343647 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkltt\" (UniqueName: \"kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt\") pod \"9c8aaf81-e916-4011-be02-a76c72a956f8\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.343704 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts\") pod \"a40cd066-4de2-427b-b1f6-207271f43d26\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.344154 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a40cd066-4de2-427b-b1f6-207271f43d26" (UID: "a40cd066-4de2-427b-b1f6-207271f43d26"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.344237 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts\") pod \"9c8aaf81-e916-4011-be02-a76c72a956f8\" (UID: \"9c8aaf81-e916-4011-be02-a76c72a956f8\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.344314 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b7nx\" (UniqueName: \"kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx\") pod \"a40cd066-4de2-427b-b1f6-207271f43d26\" (UID: \"a40cd066-4de2-427b-b1f6-207271f43d26\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.344403 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9hz8\" (UniqueName: \"kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8\") pod \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\" (UID: \"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d\") " Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.345049 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9c8aaf81-e916-4011-be02-a76c72a956f8" (UID: "9c8aaf81-e916-4011-be02-a76c72a956f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.345624 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a40cd066-4de2-427b-b1f6-207271f43d26-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.345650 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9c8aaf81-e916-4011-be02-a76c72a956f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.345659 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.345670 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.346791 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt" (OuterVolumeSpecName: "kube-api-access-vkltt") pod "9c8aaf81-e916-4011-be02-a76c72a956f8" (UID: "9c8aaf81-e916-4011-be02-a76c72a956f8"). InnerVolumeSpecName "kube-api-access-vkltt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.347717 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx" (OuterVolumeSpecName: "kube-api-access-6b7nx") pod "a40cd066-4de2-427b-b1f6-207271f43d26" (UID: "a40cd066-4de2-427b-b1f6-207271f43d26"). InnerVolumeSpecName "kube-api-access-6b7nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.348189 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8" (OuterVolumeSpecName: "kube-api-access-n9hz8") pod "7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" (UID: "7c883a5b-0f2a-48a0-b29c-9bfc01fb439d"). InnerVolumeSpecName "kube-api-access-n9hz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.348588 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l" (OuterVolumeSpecName: "kube-api-access-f5k5l") pod "79753d96-259b-4fd5-9dfe-cb27f6ce82d6" (UID: "79753d96-259b-4fd5-9dfe-cb27f6ce82d6"). InnerVolumeSpecName "kube-api-access-f5k5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.448490 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b7nx\" (UniqueName: \"kubernetes.io/projected/a40cd066-4de2-427b-b1f6-207271f43d26-kube-api-access-6b7nx\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.448540 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9hz8\" (UniqueName: \"kubernetes.io/projected/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d-kube-api-access-n9hz8\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.448552 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5k5l\" (UniqueName: \"kubernetes.io/projected/79753d96-259b-4fd5-9dfe-cb27f6ce82d6-kube-api-access-f5k5l\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.448567 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkltt\" (UniqueName: \"kubernetes.io/projected/9c8aaf81-e916-4011-be02-a76c72a956f8-kube-api-access-vkltt\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.725412 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-6f71-account-create-65cgg" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.726759 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-6f71-account-create-65cgg" event={"ID":"9c8aaf81-e916-4011-be02-a76c72a956f8","Type":"ContainerDied","Data":"018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526"} Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.726828 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="018590ffe9002a27256d83dd56467f5673a420b374601c96be3d4777837cd526" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.728284 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-a6c7-account-create-z2ldb" event={"ID":"a40cd066-4de2-427b-b1f6-207271f43d26","Type":"ContainerDied","Data":"2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93"} Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.728323 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f9903d89521f6b4580fbaba948c8894bebf13475452769ff4aa176509a70f93" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.728403 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-a6c7-account-create-z2ldb" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.745651 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b25rc" event={"ID":"7c883a5b-0f2a-48a0-b29c-9bfc01fb439d","Type":"ContainerDied","Data":"8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b"} Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.745976 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8936322ea940344a5b87a80f405b6e207414b864c63b7f63630f951637c9340b" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.745706 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b25rc" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.751271 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-vw8x4" event={"ID":"79753d96-259b-4fd5-9dfe-cb27f6ce82d6","Type":"ContainerDied","Data":"91a17f85fc5c64384c16beb71e37991af4ac850eff66b4943b0cc56fe7f9254c"} Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.751332 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91a17f85fc5c64384c16beb71e37991af4ac850eff66b4943b0cc56fe7f9254c" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.751636 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-vw8x4" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.756979 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f6ba-account-create-2c6rf" event={"ID":"23de5970-5eb9-4f95-b155-e7bc2fdbed7a","Type":"ContainerDied","Data":"d3d8664a19a5372acf79e24dabb8f691e783d348ebec05d21c408685ea8a5d18"} Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.757053 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3d8664a19a5372acf79e24dabb8f691e783d348ebec05d21c408685ea8a5d18" Nov 25 05:54:45 crc kubenswrapper[4708]: I1125 05:54:45.757064 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f6ba-account-create-2c6rf" Nov 25 05:54:46 crc kubenswrapper[4708]: I1125 05:54:46.766820 4708 generic.go:334] "Generic (PLEG): container finished" podID="cc02d86b-b016-4e10-960e-f7237575e10d" containerID="2224ff2439a37fecc13859d4d06accb76f8390714d2ebd25431b1a9ceb639735" exitCode=0 Nov 25 05:54:46 crc kubenswrapper[4708]: I1125 05:54:46.766921 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qv9pr" event={"ID":"cc02d86b-b016-4e10-960e-f7237575e10d","Type":"ContainerDied","Data":"2224ff2439a37fecc13859d4d06accb76f8390714d2ebd25431b1a9ceb639735"} Nov 25 05:54:48 crc kubenswrapper[4708]: E1125 05:54:48.471134 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.114643 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.117831 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data\") pod \"cc02d86b-b016-4e10-960e-f7237575e10d\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.117878 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxcct\" (UniqueName: \"kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct\") pod \"cc02d86b-b016-4e10-960e-f7237575e10d\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.117945 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data\") pod \"cc02d86b-b016-4e10-960e-f7237575e10d\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.117994 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle\") pod \"cc02d86b-b016-4e10-960e-f7237575e10d\" (UID: \"cc02d86b-b016-4e10-960e-f7237575e10d\") " Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.121796 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cc02d86b-b016-4e10-960e-f7237575e10d" (UID: "cc02d86b-b016-4e10-960e-f7237575e10d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.131424 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct" (OuterVolumeSpecName: "kube-api-access-jxcct") pod "cc02d86b-b016-4e10-960e-f7237575e10d" (UID: "cc02d86b-b016-4e10-960e-f7237575e10d"). InnerVolumeSpecName "kube-api-access-jxcct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.142686 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cc02d86b-b016-4e10-960e-f7237575e10d" (UID: "cc02d86b-b016-4e10-960e-f7237575e10d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.160483 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data" (OuterVolumeSpecName: "config-data") pod "cc02d86b-b016-4e10-960e-f7237575e10d" (UID: "cc02d86b-b016-4e10-960e-f7237575e10d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.220538 4708 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.220712 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.220796 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cc02d86b-b016-4e10-960e-f7237575e10d-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.220855 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxcct\" (UniqueName: \"kubernetes.io/projected/cc02d86b-b016-4e10-960e-f7237575e10d-kube-api-access-jxcct\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.801497 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wxksw" event={"ID":"31f8dae9-be16-4ee1-a485-73148c832c47","Type":"ContainerStarted","Data":"1d45c97abf29d35ddc72ef5becca9092ddd639cb53c5ca73285d0029fb1f66f0"} Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.803323 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-qv9pr" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.803294 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-qv9pr" event={"ID":"cc02d86b-b016-4e10-960e-f7237575e10d","Type":"ContainerDied","Data":"f6598c32f9a45cf357ecd9e4951db7b9cab267a391cf0c3509f717404651c8c0"} Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.803467 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6598c32f9a45cf357ecd9e4951db7b9cab267a391cf0c3509f717404651c8c0" Nov 25 05:54:49 crc kubenswrapper[4708]: I1125 05:54:49.822675 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wxksw" podStartSLOduration=6.882535159 podStartE2EDuration="13.822656132s" podCreationTimestamp="2025-11-25 05:54:36 +0000 UTC" firstStartedPulling="2025-11-25 05:54:42.104484985 +0000 UTC m=+823.513318371" lastFinishedPulling="2025-11-25 05:54:49.044605969 +0000 UTC m=+830.453439344" observedRunningTime="2025-11-25 05:54:49.817053398 +0000 UTC m=+831.225886783" watchObservedRunningTime="2025-11-25 05:54:49.822656132 +0000 UTC m=+831.231489508" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.103699 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.151378 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.151675 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="dnsmasq-dns" containerID="cri-o://1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74" gracePeriod=10 Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245360 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245757 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245777 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245802 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c33523-d685-4072-9673-9989ed7ad578" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245809 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c33523-d685-4072-9673-9989ed7ad578" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245817 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23de5970-5eb9-4f95-b155-e7bc2fdbed7a" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245823 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="23de5970-5eb9-4f95-b155-e7bc2fdbed7a" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245833 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c8aaf81-e916-4011-be02-a76c72a956f8" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245838 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c8aaf81-e916-4011-be02-a76c72a956f8" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245853 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79753d96-259b-4fd5-9dfe-cb27f6ce82d6" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245859 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="79753d96-259b-4fd5-9dfe-cb27f6ce82d6" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245869 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a40cd066-4de2-427b-b1f6-207271f43d26" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245875 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a40cd066-4de2-427b-b1f6-207271f43d26" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.245889 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc02d86b-b016-4e10-960e-f7237575e10d" containerName="glance-db-sync" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.245897 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc02d86b-b016-4e10-960e-f7237575e10d" containerName="glance-db-sync" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246193 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="23de5970-5eb9-4f95-b155-e7bc2fdbed7a" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246223 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c33523-d685-4072-9673-9989ed7ad578" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246237 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c8aaf81-e916-4011-be02-a76c72a956f8" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246246 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="79753d96-259b-4fd5-9dfe-cb27f6ce82d6" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246265 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a40cd066-4de2-427b-b1f6-207271f43d26" containerName="mariadb-account-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246273 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" containerName="mariadb-database-create" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.246305 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc02d86b-b016-4e10-960e-f7237575e10d" containerName="glance-db-sync" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.247683 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.259454 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.295266 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.112:5353: connect: connection refused" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.341553 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.341770 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzt9t\" (UniqueName: \"kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.342157 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.443063 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.443133 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzt9t\" (UniqueName: \"kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.443199 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.443722 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.443805 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.464008 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzt9t\" (UniqueName: \"kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t\") pod \"redhat-operators-g5qfx\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.498764 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.500385 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.524241 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.544946 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.545002 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-686vs\" (UniqueName: \"kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.545070 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.545119 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.545202 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.545230 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.564590 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647327 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647381 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-686vs\" (UniqueName: \"kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647417 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647446 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647501 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.647535 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.648886 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.649619 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.649808 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.649959 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.650230 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.652375 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.669981 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-686vs\" (UniqueName: \"kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs\") pod \"dnsmasq-dns-78f4d7786c-fk8jw\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.749138 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config\") pod \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.749185 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb\") pod \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.749224 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vj5xd\" (UniqueName: \"kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd\") pod \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.749286 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc\") pod \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.749392 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb\") pod \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\" (UID: \"04b54d2e-8ab3-4a86-802f-8c4abb2feb78\") " Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.774109 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd" (OuterVolumeSpecName: "kube-api-access-vj5xd") pod "04b54d2e-8ab3-4a86-802f-8c4abb2feb78" (UID: "04b54d2e-8ab3-4a86-802f-8c4abb2feb78"). InnerVolumeSpecName "kube-api-access-vj5xd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.794804 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config" (OuterVolumeSpecName: "config") pod "04b54d2e-8ab3-4a86-802f-8c4abb2feb78" (UID: "04b54d2e-8ab3-4a86-802f-8c4abb2feb78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.812124 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "04b54d2e-8ab3-4a86-802f-8c4abb2feb78" (UID: "04b54d2e-8ab3-4a86-802f-8c4abb2feb78"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.814619 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "04b54d2e-8ab3-4a86-802f-8c4abb2feb78" (UID: "04b54d2e-8ab3-4a86-802f-8c4abb2feb78"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.824496 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "04b54d2e-8ab3-4a86-802f-8c4abb2feb78" (UID: "04b54d2e-8ab3-4a86-802f-8c4abb2feb78"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.826510 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.835546 4708 generic.go:334] "Generic (PLEG): container finished" podID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerID="1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74" exitCode=0 Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.835806 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" event={"ID":"04b54d2e-8ab3-4a86-802f-8c4abb2feb78","Type":"ContainerDied","Data":"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74"} Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.835885 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" event={"ID":"04b54d2e-8ab3-4a86-802f-8c4abb2feb78","Type":"ContainerDied","Data":"87b607eaf5d85930677eb6469ae2782d6a49b6df7cc287fd73c9e57604668917"} Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.835831 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7f7d8965-gr6ck" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.835907 4708 scope.go:117] "RemoveContainer" containerID="1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.854924 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.855575 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.855593 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vj5xd\" (UniqueName: \"kubernetes.io/projected/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-kube-api-access-vj5xd\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.855608 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.855618 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/04b54d2e-8ab3-4a86-802f-8c4abb2feb78-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.867739 4708 scope.go:117] "RemoveContainer" containerID="b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.882235 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.887969 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7f7d8965-gr6ck"] Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.897625 4708 scope.go:117] "RemoveContainer" containerID="1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.899552 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74\": container with ID starting with 1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74 not found: ID does not exist" containerID="1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.899600 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74"} err="failed to get container status \"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74\": rpc error: code = NotFound desc = could not find container \"1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74\": container with ID starting with 1e7c7b913434bda42763ccc6b30834eb02c5049d1101463dc1dd154397cf9e74 not found: ID does not exist" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.899630 4708 scope.go:117] "RemoveContainer" containerID="b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4" Nov 25 05:54:50 crc kubenswrapper[4708]: E1125 05:54:50.903480 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4\": container with ID starting with b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4 not found: ID does not exist" containerID="b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.903537 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4"} err="failed to get container status \"b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4\": rpc error: code = NotFound desc = could not find container \"b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4\": container with ID starting with b0a50531ce77b9781c7e703daafa909100950c74b3d291527ac17251e35360b4 not found: ID does not exist" Nov 25 05:54:50 crc kubenswrapper[4708]: I1125 05:54:50.905440 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" path="/var/lib/kubelet/pods/04b54d2e-8ab3-4a86-802f-8c4abb2feb78/volumes" Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.015856 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.326629 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.854727 4708 generic.go:334] "Generic (PLEG): container finished" podID="8e88053d-2704-4585-906b-fd52c8d5f319" containerID="5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0" exitCode=0 Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.854806 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" event={"ID":"8e88053d-2704-4585-906b-fd52c8d5f319","Type":"ContainerDied","Data":"5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0"} Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.854865 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" event={"ID":"8e88053d-2704-4585-906b-fd52c8d5f319","Type":"ContainerStarted","Data":"ff11de0c7de8ecbccb8b1bb18dea21574893bb436a5f00953cef41f1f834a7d5"} Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.856653 4708 generic.go:334] "Generic (PLEG): container finished" podID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerID="e9e50ec12ab64ad263f86fc144d31a0837b0c7be3015e00c6b534462aadea625" exitCode=0 Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.856682 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerDied","Data":"e9e50ec12ab64ad263f86fc144d31a0837b0c7be3015e00c6b534462aadea625"} Nov 25 05:54:51 crc kubenswrapper[4708]: I1125 05:54:51.856712 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerStarted","Data":"32c448398b7f7c739c9966fdcb1471f4ae5c5e20570a1e0b10c679b5debbd407"} Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.868971 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerStarted","Data":"de1fdaa727feb2782a730140b3db205f7dca6d9fabf5e9836dcc5e54e34ec73f"} Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.872138 4708 generic.go:334] "Generic (PLEG): container finished" podID="31f8dae9-be16-4ee1-a485-73148c832c47" containerID="1d45c97abf29d35ddc72ef5becca9092ddd639cb53c5ca73285d0029fb1f66f0" exitCode=0 Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.872204 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wxksw" event={"ID":"31f8dae9-be16-4ee1-a485-73148c832c47","Type":"ContainerDied","Data":"1d45c97abf29d35ddc72ef5becca9092ddd639cb53c5ca73285d0029fb1f66f0"} Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.874746 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" event={"ID":"8e88053d-2704-4585-906b-fd52c8d5f319","Type":"ContainerStarted","Data":"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f"} Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.875358 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:52 crc kubenswrapper[4708]: I1125 05:54:52.928285 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" podStartSLOduration=2.928264066 podStartE2EDuration="2.928264066s" podCreationTimestamp="2025-11-25 05:54:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:52.920783273 +0000 UTC m=+834.329616660" watchObservedRunningTime="2025-11-25 05:54:52.928264066 +0000 UTC m=+834.337097451" Nov 25 05:54:53 crc kubenswrapper[4708]: I1125 05:54:53.884966 4708 generic.go:334] "Generic (PLEG): container finished" podID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerID="de1fdaa727feb2782a730140b3db205f7dca6d9fabf5e9836dcc5e54e34ec73f" exitCode=0 Nov 25 05:54:53 crc kubenswrapper[4708]: I1125 05:54:53.885066 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerDied","Data":"de1fdaa727feb2782a730140b3db205f7dca6d9fabf5e9836dcc5e54e34ec73f"} Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.168732 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.326338 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle\") pod \"31f8dae9-be16-4ee1-a485-73148c832c47\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.326480 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4dbl\" (UniqueName: \"kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl\") pod \"31f8dae9-be16-4ee1-a485-73148c832c47\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.326512 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data\") pod \"31f8dae9-be16-4ee1-a485-73148c832c47\" (UID: \"31f8dae9-be16-4ee1-a485-73148c832c47\") " Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.333208 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl" (OuterVolumeSpecName: "kube-api-access-j4dbl") pod "31f8dae9-be16-4ee1-a485-73148c832c47" (UID: "31f8dae9-be16-4ee1-a485-73148c832c47"). InnerVolumeSpecName "kube-api-access-j4dbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.349789 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31f8dae9-be16-4ee1-a485-73148c832c47" (UID: "31f8dae9-be16-4ee1-a485-73148c832c47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.379050 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data" (OuterVolumeSpecName: "config-data") pod "31f8dae9-be16-4ee1-a485-73148c832c47" (UID: "31f8dae9-be16-4ee1-a485-73148c832c47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.428461 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.428497 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4dbl\" (UniqueName: \"kubernetes.io/projected/31f8dae9-be16-4ee1-a485-73148c832c47-kube-api-access-j4dbl\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.428509 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31f8dae9-be16-4ee1-a485-73148c832c47-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.903935 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wxksw" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.907393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerStarted","Data":"090cb4f191a00060e121de317f8c3f7cd94aa6006003adf807770b67f42e804a"} Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.907462 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wxksw" event={"ID":"31f8dae9-be16-4ee1-a485-73148c832c47","Type":"ContainerDied","Data":"d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b"} Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.907482 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d09ce9074aaafc06b9cb1164744028e02d3784008620e0614bba81f95e27316b" Nov 25 05:54:54 crc kubenswrapper[4708]: I1125 05:54:54.939797 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5qfx" podStartSLOduration=2.408404913 podStartE2EDuration="4.939777163s" podCreationTimestamp="2025-11-25 05:54:50 +0000 UTC" firstStartedPulling="2025-11-25 05:54:51.862244466 +0000 UTC m=+833.271077841" lastFinishedPulling="2025-11-25 05:54:54.393616705 +0000 UTC m=+835.802450091" observedRunningTime="2025-11-25 05:54:54.929209906 +0000 UTC m=+836.338043293" watchObservedRunningTime="2025-11-25 05:54:54.939777163 +0000 UTC m=+836.348610548" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.164859 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.193728 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-q924z"] Nov 25 05:54:55 crc kubenswrapper[4708]: E1125 05:54:55.194094 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="dnsmasq-dns" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.194115 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="dnsmasq-dns" Nov 25 05:54:55 crc kubenswrapper[4708]: E1125 05:54:55.194128 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f8dae9-be16-4ee1-a485-73148c832c47" containerName="keystone-db-sync" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.194135 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f8dae9-be16-4ee1-a485-73148c832c47" containerName="keystone-db-sync" Nov 25 05:54:55 crc kubenswrapper[4708]: E1125 05:54:55.194159 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="init" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.194165 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="init" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.194397 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f8dae9-be16-4ee1-a485-73148c832c47" containerName="keystone-db-sync" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.194412 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="04b54d2e-8ab3-4a86-802f-8c4abb2feb78" containerName="dnsmasq-dns" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.195006 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.197151 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.197192 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.197406 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.197904 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sdz4c" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.205173 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-759c677775-gk2wh"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.207485 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.208854 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.223026 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q924z"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.230639 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-759c677775-gk2wh"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.352916 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.352970 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.352991 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2wxq\" (UniqueName: \"kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353019 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353052 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353078 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gpnn\" (UniqueName: \"kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353098 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353129 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353146 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353179 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353232 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.353268 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.358497 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.363100 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.369970 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.370177 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.378459 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.413465 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-gnqch"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.414425 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.420692 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.420917 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.421033 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-np7px" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.444512 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gnqch"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455589 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455650 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455679 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455720 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455771 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455851 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455889 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455923 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455951 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455978 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.455999 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2wxq\" (UniqueName: \"kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456022 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456044 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456075 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwvrl\" (UniqueName: \"kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456095 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456123 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456144 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gpnn\" (UniqueName: \"kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.456163 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.457164 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.457253 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.457892 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.459965 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.462288 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-vcmh6"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.463544 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.470137 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n6chh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.470413 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.476233 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.480921 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481077 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481168 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbk69\" (UniqueName: \"kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481215 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481246 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481311 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481339 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481391 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481407 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.481508 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.482359 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.482979 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2wxq\" (UniqueName: \"kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.487507 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vcmh6"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.498440 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759c677775-gk2wh"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.506044 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.507244 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data\") pod \"keystone-bootstrap-q924z\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: E1125 05:54:55.511239 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-7gpnn], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-759c677775-gk2wh" podUID="d0787713-ee34-4006-bce7-a6c23ddf8ed2" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.511864 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q924z" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.512861 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gpnn\" (UniqueName: \"kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn\") pod \"dnsmasq-dns-759c677775-gk2wh\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584674 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584704 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584722 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584743 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584765 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584785 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpktm\" (UniqueName: \"kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584811 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584837 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwvrl\" (UniqueName: \"kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584860 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584884 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbk69\" (UniqueName: \"kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584915 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584942 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584960 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.584988 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.585011 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.585182 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.585882 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.590019 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.590097 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.590959 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.592170 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.594243 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.595647 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.595968 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cqjrf"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.597480 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.599972 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.600479 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.601628 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.601958 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s5nlv" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.602155 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.606187 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.608859 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.610739 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.614136 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwvrl\" (UniqueName: \"kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl\") pod \"cinder-db-sync-gnqch\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.614498 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbk69\" (UniqueName: \"kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69\") pod \"ceilometer-0\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.636529 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cqjrf"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.659942 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-lgvsg"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.660761 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.662994 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-lgvsg"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.664884 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.664913 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dhz8l" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.673903 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.682052 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687191 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687244 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687345 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687380 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687402 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687443 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpktm\" (UniqueName: \"kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687482 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687545 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687589 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rhx7\" (UniqueName: \"kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687617 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687659 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687701 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687718 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.687752 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grcbg\" (UniqueName: \"kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.692340 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.697640 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.707105 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpktm\" (UniqueName: \"kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm\") pod \"neutron-db-sync-vcmh6\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.742067 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gnqch" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789465 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789555 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789601 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rhx7\" (UniqueName: \"kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789622 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789660 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789685 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789703 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789732 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grcbg\" (UniqueName: \"kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789751 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789779 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789802 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789876 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789928 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.789957 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcxp7\" (UniqueName: \"kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.791006 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.791681 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.792422 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.793678 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.793953 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.794464 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.797018 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.802213 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.804476 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.838852 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grcbg\" (UniqueName: \"kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg\") pod \"placement-db-sync-cqjrf\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.852053 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rhx7\" (UniqueName: \"kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7\") pod \"dnsmasq-dns-69d79494c5-xf4kn\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.873964 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.892209 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcxp7\" (UniqueName: \"kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.892332 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.898937 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.923075 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.940315 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cqjrf" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.944731 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.945992 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="dnsmasq-dns" containerID="cri-o://215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f" gracePeriod=10 Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.948312 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.950208 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcxp7\" (UniqueName: \"kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.951479 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data\") pod \"barbican-db-sync-lgvsg\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:55 crc kubenswrapper[4708]: I1125 05:54:55.989785 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.030456 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.075867 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:54:56 crc kubenswrapper[4708]: W1125 05:54:56.100323 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd07de339_30ce_4edc_86f5_2f2dcebc417e.slice/crio-5c1b127e10d58661b19d328167cfc361999745ee6dad5810a50d3569e9d887b8 WatchSource:0}: Error finding container 5c1b127e10d58661b19d328167cfc361999745ee6dad5810a50d3569e9d887b8: Status 404 returned error can't find the container with id 5c1b127e10d58661b19d328167cfc361999745ee6dad5810a50d3569e9d887b8 Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106260 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106388 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gpnn\" (UniqueName: \"kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106500 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106553 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106607 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.106698 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb\") pod \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\" (UID: \"d0787713-ee34-4006-bce7-a6c23ddf8ed2\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.108364 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config" (OuterVolumeSpecName: "config") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.108450 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.108938 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.109165 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.110936 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.119287 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn" (OuterVolumeSpecName: "kube-api-access-7gpnn") pod "d0787713-ee34-4006-bce7-a6c23ddf8ed2" (UID: "d0787713-ee34-4006-bce7-a6c23ddf8ed2"). InnerVolumeSpecName "kube-api-access-7gpnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.138683 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-q924z"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209680 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209701 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209714 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209724 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209733 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0787713-ee34-4006-bce7-a6c23ddf8ed2-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.209740 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gpnn\" (UniqueName: \"kubernetes.io/projected/d0787713-ee34-4006-bce7-a6c23ddf8ed2-kube-api-access-7gpnn\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.307249 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.309101 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.312688 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.312975 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.313368 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-j2dg4" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.313651 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.315708 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.345451 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.353846 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.362095 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.362617 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.379163 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416199 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416237 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j98mg\" (UniqueName: \"kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416263 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416333 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416353 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416403 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsj25\" (UniqueName: \"kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416428 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416480 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416562 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416593 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416642 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416932 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416955 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.416972 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.417146 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.417198 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.429489 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-gnqch"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.488432 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.520824 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.520901 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.520975 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.521074 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-686vs\" (UniqueName: \"kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.521171 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.521263 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc\") pod \"8e88053d-2704-4585-906b-fd52c8d5f319\" (UID: \"8e88053d-2704-4585-906b-fd52c8d5f319\") " Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522144 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522179 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522228 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522329 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522385 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522410 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522478 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522534 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522609 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522640 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j98mg\" (UniqueName: \"kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522674 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522703 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522722 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522750 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsj25\" (UniqueName: \"kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522784 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.522820 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.525060 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.529357 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.529967 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.534771 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.535789 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.535934 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs" (OuterVolumeSpecName: "kube-api-access-686vs") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "kube-api-access-686vs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.537005 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.537401 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.537732 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.538165 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.548808 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.549059 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.553119 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.553685 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.553868 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.556045 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j98mg\" (UniqueName: \"kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.561384 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsj25\" (UniqueName: \"kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.598360 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.604128 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.607947 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config" (OuterVolumeSpecName: "config") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.615168 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.620874 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.633146 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-686vs\" (UniqueName: \"kubernetes.io/projected/8e88053d-2704-4585-906b-fd52c8d5f319-kube-api-access-686vs\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.633174 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.633189 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.633199 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.635906 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.640815 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e88053d-2704-4585-906b-fd52c8d5f319" (UID: "8e88053d-2704-4585-906b-fd52c8d5f319"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.644099 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.659775 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:54:56 crc kubenswrapper[4708]: E1125 05:54:56.660235 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="dnsmasq-dns" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.660253 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="dnsmasq-dns" Nov 25 05:54:56 crc kubenswrapper[4708]: E1125 05:54:56.660281 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="init" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.660289 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="init" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.660483 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" containerName="dnsmasq-dns" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.666854 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-vcmh6"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.666965 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.670503 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.684999 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.735074 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxkm9\" (UniqueName: \"kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.735631 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.735672 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.735728 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.735739 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e88053d-2704-4585-906b-fd52c8d5f319-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.752616 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.758419 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cqjrf"] Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.774663 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-lgvsg"] Nov 25 05:54:56 crc kubenswrapper[4708]: W1125 05:54:56.791579 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f0e677f_bc35_4c02_8322_37d4cf7fc350.slice/crio-bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a WatchSource:0}: Error finding container bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a: Status 404 returned error can't find the container with id bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a Nov 25 05:54:56 crc kubenswrapper[4708]: W1125 05:54:56.797805 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod454358fb_f82d_4c19_b6e3_5f1176829d6e.slice/crio-02efa37391ee85e114fa802e9c2fd2b53ccec4638929bbe55422a1011bc86a0e WatchSource:0}: Error finding container 02efa37391ee85e114fa802e9c2fd2b53ccec4638929bbe55422a1011bc86a0e: Status 404 returned error can't find the container with id 02efa37391ee85e114fa802e9c2fd2b53ccec4638929bbe55422a1011bc86a0e Nov 25 05:54:56 crc kubenswrapper[4708]: W1125 05:54:56.816209 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5726473d_059a_4f7d_ac6b_4762c732f8ff.slice/crio-63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b WatchSource:0}: Error finding container 63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b: Status 404 returned error can't find the container with id 63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.836931 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.836966 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.836995 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxkm9\" (UniqueName: \"kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.837675 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.837885 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.859785 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxkm9\" (UniqueName: \"kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9\") pod \"certified-operators-2qhfv\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.972448 4708 generic.go:334] "Generic (PLEG): container finished" podID="8e88053d-2704-4585-906b-fd52c8d5f319" containerID="215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f" exitCode=0 Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.972802 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" event={"ID":"8e88053d-2704-4585-906b-fd52c8d5f319","Type":"ContainerDied","Data":"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f"} Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.972834 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" event={"ID":"8e88053d-2704-4585-906b-fd52c8d5f319","Type":"ContainerDied","Data":"ff11de0c7de8ecbccb8b1bb18dea21574893bb436a5f00953cef41f1f834a7d5"} Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.972879 4708 scope.go:117] "RemoveContainer" containerID="215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.974479 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerStarted","Data":"5c1b127e10d58661b19d328167cfc361999745ee6dad5810a50d3569e9d887b8"} Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.988099 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78f4d7786c-fk8jw" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.989777 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.997770 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q924z" event={"ID":"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c","Type":"ContainerStarted","Data":"abbafe6a07a038f715d82b1a8caf876dcca2d534627ad779c64b17e6e5ac4455"} Nov 25 05:54:56 crc kubenswrapper[4708]: I1125 05:54:56.997828 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q924z" event={"ID":"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c","Type":"ContainerStarted","Data":"c6acf8e8bf3f1b4f8a01be26886bcef6e09439de6dd72f0a1be838f5e7f8164f"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.001408 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" event={"ID":"454358fb-f82d-4c19-b6e3-5f1176829d6e","Type":"ContainerStarted","Data":"02efa37391ee85e114fa802e9c2fd2b53ccec4638929bbe55422a1011bc86a0e"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.003673 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gnqch" event={"ID":"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f","Type":"ContainerStarted","Data":"de1838a4044bddba3eb99f6a88a12f1627d70b4384f7b1d577fd41adbe7f54a3"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.022442 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vcmh6" event={"ID":"9174a689-38fd-478f-8c3e-07ab7e21a921","Type":"ContainerStarted","Data":"49fc37dedeb9d41de6a45b462c37c6897c703a80b40d217ec187796af48c2362"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.025971 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lgvsg" event={"ID":"0f0e677f-bc35-4c02-8322-37d4cf7fc350","Type":"ContainerStarted","Data":"bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.026583 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cqjrf" event={"ID":"5726473d-059a-4f7d-ac6b-4762c732f8ff","Type":"ContainerStarted","Data":"63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b"} Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.025192 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-759c677775-gk2wh" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.030146 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.051164 4708 scope.go:117] "RemoveContainer" containerID="5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.052824 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78f4d7786c-fk8jw"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.063222 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-q924z" podStartSLOduration=2.063206804 podStartE2EDuration="2.063206804s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:57.029274493 +0000 UTC m=+838.438107879" watchObservedRunningTime="2025-11-25 05:54:57.063206804 +0000 UTC m=+838.472040190" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.136489 4708 scope.go:117] "RemoveContainer" containerID="215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f" Nov 25 05:54:57 crc kubenswrapper[4708]: E1125 05:54:57.137969 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f\": container with ID starting with 215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f not found: ID does not exist" containerID="215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.138005 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f"} err="failed to get container status \"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f\": rpc error: code = NotFound desc = could not find container \"215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f\": container with ID starting with 215ca492b3f8eb4e9b5bf1873a8463ec94fcd4ed1c7109d296b79b2fffc8223f not found: ID does not exist" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.138032 4708 scope.go:117] "RemoveContainer" containerID="5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0" Nov 25 05:54:57 crc kubenswrapper[4708]: E1125 05:54:57.138450 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0\": container with ID starting with 5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0 not found: ID does not exist" containerID="5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.138476 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0"} err="failed to get container status \"5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0\": rpc error: code = NotFound desc = could not find container \"5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0\": container with ID starting with 5d839b110ed20e1d6931a70a085b3d2c9b7cd7dfe3fd323b916d6d4e1e7c36a0 not found: ID does not exist" Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.170930 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-759c677775-gk2wh"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.188294 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-759c677775-gk2wh"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.258530 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.345984 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.404017 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.413457 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.458488 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:54:57 crc kubenswrapper[4708]: I1125 05:54:57.657460 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:54:57 crc kubenswrapper[4708]: W1125 05:54:57.684758 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd244e370_f811_42f5_9036_7ce6895289ab.slice/crio-380fe082bb5cbddcc491b7db291801695e0632e94a032027780e80c78c1307a9 WatchSource:0}: Error finding container 380fe082bb5cbddcc491b7db291801695e0632e94a032027780e80c78c1307a9: Status 404 returned error can't find the container with id 380fe082bb5cbddcc491b7db291801695e0632e94a032027780e80c78c1307a9 Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.047779 4708 generic.go:334] "Generic (PLEG): container finished" podID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerID="72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca" exitCode=0 Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.048062 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" event={"ID":"454358fb-f82d-4c19-b6e3-5f1176829d6e","Type":"ContainerDied","Data":"72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca"} Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.050406 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerStarted","Data":"d067f8cdd33511677f262fa65d5e301ca7d995bb4a1188ccea3fd5db2d93196b"} Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.081104 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vcmh6" event={"ID":"9174a689-38fd-478f-8c3e-07ab7e21a921","Type":"ContainerStarted","Data":"3977e9d7662fa5bf82ec5c52c5415bf8e0eb59e913711fe590c89f19eae60060"} Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.122360 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-vcmh6" podStartSLOduration=3.122339951 podStartE2EDuration="3.122339951s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:54:58.113637208 +0000 UTC m=+839.522470594" watchObservedRunningTime="2025-11-25 05:54:58.122339951 +0000 UTC m=+839.531173337" Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.132215 4708 generic.go:334] "Generic (PLEG): container finished" podID="d244e370-f811-42f5-9036-7ce6895289ab" containerID="504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c" exitCode=0 Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.132280 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerDied","Data":"504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c"} Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.132318 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerStarted","Data":"380fe082bb5cbddcc491b7db291801695e0632e94a032027780e80c78c1307a9"} Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.180198 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerStarted","Data":"a25e5a2508e1c94095c13be19e7074aa92567c726b016a43939543fd87d03ffa"} Nov 25 05:54:58 crc kubenswrapper[4708]: E1125 05:54:58.753922 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01fd2cf9_a2a2_4fb3_ab9e_1d60fea51ecf.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.928869 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e88053d-2704-4585-906b-fd52c8d5f319" path="/var/lib/kubelet/pods/8e88053d-2704-4585-906b-fd52c8d5f319/volumes" Nov 25 05:54:58 crc kubenswrapper[4708]: I1125 05:54:58.929506 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0787713-ee34-4006-bce7-a6c23ddf8ed2" path="/var/lib/kubelet/pods/d0787713-ee34-4006-bce7-a6c23ddf8ed2/volumes" Nov 25 05:54:59 crc kubenswrapper[4708]: I1125 05:54:59.223218 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerStarted","Data":"c4b97ef4e5af2850fbd1c6e7912f5404dded7fd308785bf0b12be940cd5ffa35"} Nov 25 05:54:59 crc kubenswrapper[4708]: I1125 05:54:59.230713 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" event={"ID":"454358fb-f82d-4c19-b6e3-5f1176829d6e","Type":"ContainerStarted","Data":"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd"} Nov 25 05:54:59 crc kubenswrapper[4708]: I1125 05:54:59.239027 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerStarted","Data":"ad9f4ec7c5b2d475f3f0798766a0daaf1ff15dd567543af775c8b6885469b697"} Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.251139 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerStarted","Data":"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83"} Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.253848 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerStarted","Data":"f960cc33ea47dc79cf61a83814f0b8f718685a43ff98f0acaa071283b85d9daf"} Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.253988 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-log" containerID="cri-o://c4b97ef4e5af2850fbd1c6e7912f5404dded7fd308785bf0b12be940cd5ffa35" gracePeriod=30 Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.254213 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-httpd" containerID="cri-o://f960cc33ea47dc79cf61a83814f0b8f718685a43ff98f0acaa071283b85d9daf" gracePeriod=30 Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.257723 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-log" containerID="cri-o://ad9f4ec7c5b2d475f3f0798766a0daaf1ff15dd567543af775c8b6885469b697" gracePeriod=30 Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.257929 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerStarted","Data":"80c44d8485d61922521148228a03401a73cb6b11ad28fb05c8b9c436168397db"} Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.257982 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.257931 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-httpd" containerID="cri-o://80c44d8485d61922521148228a03401a73cb6b11ad28fb05c8b9c436168397db" gracePeriod=30 Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.300569 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.300558543 podStartE2EDuration="5.300558543s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:00.294852323 +0000 UTC m=+841.703685709" watchObservedRunningTime="2025-11-25 05:55:00.300558543 +0000 UTC m=+841.709391929" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.350543 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" podStartSLOduration=5.350511211 podStartE2EDuration="5.350511211s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:00.347326032 +0000 UTC m=+841.756159417" watchObservedRunningTime="2025-11-25 05:55:00.350511211 +0000 UTC m=+841.759344597" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.384295 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.384270696 podStartE2EDuration="5.384270696s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:00.382259017 +0000 UTC m=+841.791092403" watchObservedRunningTime="2025-11-25 05:55:00.384270696 +0000 UTC m=+841.793104082" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.565586 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.566654 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:00 crc kubenswrapper[4708]: I1125 05:55:00.636170 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.270093 4708 generic.go:334] "Generic (PLEG): container finished" podID="d244e370-f811-42f5-9036-7ce6895289ab" containerID="93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83" exitCode=0 Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.270151 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerDied","Data":"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83"} Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.274709 4708 generic.go:334] "Generic (PLEG): container finished" podID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerID="f960cc33ea47dc79cf61a83814f0b8f718685a43ff98f0acaa071283b85d9daf" exitCode=0 Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.274942 4708 generic.go:334] "Generic (PLEG): container finished" podID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerID="c4b97ef4e5af2850fbd1c6e7912f5404dded7fd308785bf0b12be940cd5ffa35" exitCode=143 Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.274993 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerDied","Data":"f960cc33ea47dc79cf61a83814f0b8f718685a43ff98f0acaa071283b85d9daf"} Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.275020 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerDied","Data":"c4b97ef4e5af2850fbd1c6e7912f5404dded7fd308785bf0b12be940cd5ffa35"} Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.278787 4708 generic.go:334] "Generic (PLEG): container finished" podID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerID="80c44d8485d61922521148228a03401a73cb6b11ad28fb05c8b9c436168397db" exitCode=0 Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.279374 4708 generic.go:334] "Generic (PLEG): container finished" podID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerID="ad9f4ec7c5b2d475f3f0798766a0daaf1ff15dd567543af775c8b6885469b697" exitCode=143 Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.279180 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerDied","Data":"80c44d8485d61922521148228a03401a73cb6b11ad28fb05c8b9c436168397db"} Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.279462 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerDied","Data":"ad9f4ec7c5b2d475f3f0798766a0daaf1ff15dd567543af775c8b6885469b697"} Nov 25 05:55:01 crc kubenswrapper[4708]: I1125 05:55:01.336869 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:02 crc kubenswrapper[4708]: I1125 05:55:02.290378 4708 generic.go:334] "Generic (PLEG): container finished" podID="c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" containerID="abbafe6a07a038f715d82b1a8caf876dcca2d534627ad779c64b17e6e5ac4455" exitCode=0 Nov 25 05:55:02 crc kubenswrapper[4708]: I1125 05:55:02.290506 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q924z" event={"ID":"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c","Type":"ContainerDied","Data":"abbafe6a07a038f715d82b1a8caf876dcca2d534627ad779c64b17e6e5ac4455"} Nov 25 05:55:02 crc kubenswrapper[4708]: I1125 05:55:02.624700 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.245896 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.249744 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.252653 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.341912 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dpxl\" (UniqueName: \"kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.342050 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.342110 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.446322 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.446549 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.446717 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dpxl\" (UniqueName: \"kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.446925 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.447171 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.468232 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dpxl\" (UniqueName: \"kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl\") pod \"redhat-marketplace-srn5s\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:03 crc kubenswrapper[4708]: I1125 05:55:03.570952 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.093185 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q924z" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162451 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162629 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162660 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162713 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162878 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2wxq\" (UniqueName: \"kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.162902 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle\") pod \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\" (UID: \"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c\") " Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.167790 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq" (OuterVolumeSpecName: "kube-api-access-s2wxq") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "kube-api-access-s2wxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.173936 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts" (OuterVolumeSpecName: "scripts") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.174568 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.177188 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.190031 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data" (OuterVolumeSpecName: "config-data") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.202574 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" (UID: "c88871b8-2ff6-4cfa-aa1b-3121dc75be1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265296 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2wxq\" (UniqueName: \"kubernetes.io/projected/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-kube-api-access-s2wxq\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265572 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265585 4708 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265597 4708 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265606 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.265615 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.333715 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-q924z" event={"ID":"c88871b8-2ff6-4cfa-aa1b-3121dc75be1c","Type":"ContainerDied","Data":"c6acf8e8bf3f1b4f8a01be26886bcef6e09439de6dd72f0a1be838f5e7f8164f"} Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.333785 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6acf8e8bf3f1b4f8a01be26886bcef6e09439de6dd72f0a1be838f5e7f8164f" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.333784 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-q924z" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.333880 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5qfx" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="registry-server" containerID="cri-o://090cb4f191a00060e121de317f8c3f7cd94aa6006003adf807770b67f42e804a" gracePeriod=2 Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.380087 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-q924z"] Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.387897 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-q924z"] Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.464888 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8kfbn"] Nov 25 05:55:04 crc kubenswrapper[4708]: E1125 05:55:04.465376 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" containerName="keystone-bootstrap" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.465399 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" containerName="keystone-bootstrap" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.465613 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" containerName="keystone-bootstrap" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.466244 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470199 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470405 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8kfbn"] Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470460 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470601 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sdz4c" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470721 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.470948 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574190 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574251 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574317 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574345 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574416 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.574435 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhr52\" (UniqueName: \"kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.676215 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.676268 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.676340 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.676359 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.676999 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.677023 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhr52\" (UniqueName: \"kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.681049 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.681835 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.682016 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.682329 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.688611 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.690491 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhr52\" (UniqueName: \"kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52\") pod \"keystone-bootstrap-8kfbn\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.809371 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:04 crc kubenswrapper[4708]: I1125 05:55:04.909462 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c88871b8-2ff6-4cfa-aa1b-3121dc75be1c" path="/var/lib/kubelet/pods/c88871b8-2ff6-4cfa-aa1b-3121dc75be1c/volumes" Nov 25 05:55:05 crc kubenswrapper[4708]: I1125 05:55:05.366192 4708 generic.go:334] "Generic (PLEG): container finished" podID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerID="090cb4f191a00060e121de317f8c3f7cd94aa6006003adf807770b67f42e804a" exitCode=0 Nov 25 05:55:05 crc kubenswrapper[4708]: I1125 05:55:05.366295 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerDied","Data":"090cb4f191a00060e121de317f8c3f7cd94aa6006003adf807770b67f42e804a"} Nov 25 05:55:05 crc kubenswrapper[4708]: I1125 05:55:05.949676 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.000652 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.000922 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="dnsmasq-dns" containerID="cri-o://f38ca8faa43e41d183bca58a0da7dd333b2abd599cdad0cc03e37b376b19542d" gracePeriod=10 Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.162814 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.212450 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.212594 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.212859 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsj25\" (UniqueName: \"kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.212918 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.212990 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.213011 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.213069 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.213131 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts\") pod \"eec5ab77-2772-4e84-9303-67a41dd59c07\" (UID: \"eec5ab77-2772-4e84-9303-67a41dd59c07\") " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.213803 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.213900 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs" (OuterVolumeSpecName: "logs") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.214289 4708 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.214317 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/eec5ab77-2772-4e84-9303-67a41dd59c07-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.273252 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.291218 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25" (OuterVolumeSpecName: "kube-api-access-wsj25") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "kube-api-access-wsj25". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.291788 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts" (OuterVolumeSpecName: "scripts") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.320898 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.320934 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsj25\" (UniqueName: \"kubernetes.io/projected/eec5ab77-2772-4e84-9303-67a41dd59c07-kube-api-access-wsj25\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.320960 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.360017 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.363198 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.380317 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data" (OuterVolumeSpecName: "config-data") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.382194 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"eec5ab77-2772-4e84-9303-67a41dd59c07","Type":"ContainerDied","Data":"d067f8cdd33511677f262fa65d5e301ca7d995bb4a1188ccea3fd5db2d93196b"} Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.382247 4708 scope.go:117] "RemoveContainer" containerID="80c44d8485d61922521148228a03401a73cb6b11ad28fb05c8b9c436168397db" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.382383 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.389442 4708 generic.go:334] "Generic (PLEG): container finished" podID="9174a689-38fd-478f-8c3e-07ab7e21a921" containerID="3977e9d7662fa5bf82ec5c52c5415bf8e0eb59e913711fe590c89f19eae60060" exitCode=0 Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.389490 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vcmh6" event={"ID":"9174a689-38fd-478f-8c3e-07ab7e21a921","Type":"ContainerDied","Data":"3977e9d7662fa5bf82ec5c52c5415bf8e0eb59e913711fe590c89f19eae60060"} Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.395617 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "eec5ab77-2772-4e84-9303-67a41dd59c07" (UID: "eec5ab77-2772-4e84-9303-67a41dd59c07"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.400761 4708 generic.go:334] "Generic (PLEG): container finished" podID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerID="f38ca8faa43e41d183bca58a0da7dd333b2abd599cdad0cc03e37b376b19542d" exitCode=0 Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.400806 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" event={"ID":"59c87445-1528-4cab-9a2a-8d736c1a6c71","Type":"ContainerDied","Data":"f38ca8faa43e41d183bca58a0da7dd333b2abd599cdad0cc03e37b376b19542d"} Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.423285 4708 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.423323 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.423334 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.423348 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eec5ab77-2772-4e84-9303-67a41dd59c07-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.737940 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.753926 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.767262 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:55:06 crc kubenswrapper[4708]: E1125 05:55:06.767973 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-httpd" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.767998 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-httpd" Nov 25 05:55:06 crc kubenswrapper[4708]: E1125 05:55:06.768011 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-log" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.768018 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-log" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.768245 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-log" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.768272 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" containerName="glance-httpd" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.769822 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.771895 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.772255 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.777074 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835423 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835496 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835537 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75p4k\" (UniqueName: \"kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835582 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835601 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835637 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.835666 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.903729 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eec5ab77-2772-4e84-9303-67a41dd59c07" path="/var/lib/kubelet/pods/eec5ab77-2772-4e84-9303-67a41dd59c07/volumes" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.936982 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937028 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937109 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937159 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937206 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75p4k\" (UniqueName: \"kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937231 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937252 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937270 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.937589 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.938155 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.938180 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.941504 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.941837 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.942016 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.942436 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.959541 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:06 crc kubenswrapper[4708]: I1125 05:55:06.962939 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75p4k\" (UniqueName: \"kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k\") pod \"glance-default-external-api-0\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " pod="openstack/glance-default-external-api-0" Nov 25 05:55:07 crc kubenswrapper[4708]: I1125 05:55:07.096173 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.418497 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"04465ec9-6070-4ec3-89e0-c45b84056e60","Type":"ContainerDied","Data":"a25e5a2508e1c94095c13be19e7074aa92567c726b016a43939543fd87d03ffa"} Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.418935 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a25e5a2508e1c94095c13be19e7074aa92567c726b016a43939543fd87d03ffa" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.486616 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.490011 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568277 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568358 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config\") pod \"9174a689-38fd-478f-8c3e-07ab7e21a921\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568412 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568439 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpktm\" (UniqueName: \"kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm\") pod \"9174a689-38fd-478f-8c3e-07ab7e21a921\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568531 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j98mg\" (UniqueName: \"kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568622 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568703 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568732 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568749 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle\") pod \"9174a689-38fd-478f-8c3e-07ab7e21a921\" (UID: \"9174a689-38fd-478f-8c3e-07ab7e21a921\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568812 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.568837 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs\") pod \"04465ec9-6070-4ec3-89e0-c45b84056e60\" (UID: \"04465ec9-6070-4ec3-89e0-c45b84056e60\") " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.570508 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs" (OuterVolumeSpecName: "logs") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.584902 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.593637 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.601628 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm" (OuterVolumeSpecName: "kube-api-access-tpktm") pod "9174a689-38fd-478f-8c3e-07ab7e21a921" (UID: "9174a689-38fd-478f-8c3e-07ab7e21a921"). InnerVolumeSpecName "kube-api-access-tpktm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.609742 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts" (OuterVolumeSpecName: "scripts") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.629658 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg" (OuterVolumeSpecName: "kube-api-access-j98mg") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "kube-api-access-j98mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671034 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j98mg\" (UniqueName: \"kubernetes.io/projected/04465ec9-6070-4ec3-89e0-c45b84056e60-kube-api-access-j98mg\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671080 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671090 4708 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671098 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04465ec9-6070-4ec3-89e0-c45b84056e60-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671107 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.671116 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpktm\" (UniqueName: \"kubernetes.io/projected/9174a689-38fd-478f-8c3e-07ab7e21a921-kube-api-access-tpktm\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.688675 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9174a689-38fd-478f-8c3e-07ab7e21a921" (UID: "9174a689-38fd-478f-8c3e-07ab7e21a921"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.696145 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.715673 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config" (OuterVolumeSpecName: "config") pod "9174a689-38fd-478f-8c3e-07ab7e21a921" (UID: "9174a689-38fd-478f-8c3e-07ab7e21a921"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.733827 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.769377 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data" (OuterVolumeSpecName: "config-data") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.771978 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.772010 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.772019 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.772029 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.772038 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9174a689-38fd-478f-8c3e-07ab7e21a921-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.772586 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "04465ec9-6070-4ec3-89e0-c45b84056e60" (UID: "04465ec9-6070-4ec3-89e0-c45b84056e60"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:08 crc kubenswrapper[4708]: I1125 05:55:08.873779 4708 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04465ec9-6070-4ec3-89e0-c45b84056e60-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.001121 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod04465ec9_6070_4ec3_89e0_c45b84056e60.slice/crio-a25e5a2508e1c94095c13be19e7074aa92567c726b016a43939543fd87d03ffa\": RecentStats: unable to find data in memory cache]" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.120049 4708 scope.go:117] "RemoveContainer" containerID="ad9f4ec7c5b2d475f3f0798766a0daaf1ff15dd567543af775c8b6885469b697" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.136590 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.141882 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185103 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzt9t\" (UniqueName: \"kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t\") pod \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185212 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities\") pod \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185266 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g666h\" (UniqueName: \"kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185286 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185340 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185388 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185435 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185468 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0\") pod \"59c87445-1528-4cab-9a2a-8d736c1a6c71\" (UID: \"59c87445-1528-4cab-9a2a-8d736c1a6c71\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.185600 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content\") pod \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\" (UID: \"c43835f5-9b6e-4b72-8c1d-0bcb32230845\") " Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.187792 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities" (OuterVolumeSpecName: "utilities") pod "c43835f5-9b6e-4b72-8c1d-0bcb32230845" (UID: "c43835f5-9b6e-4b72-8c1d-0bcb32230845"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.193010 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.193402 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h" (OuterVolumeSpecName: "kube-api-access-g666h") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "kube-api-access-g666h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.195418 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t" (OuterVolumeSpecName: "kube-api-access-gzt9t") pod "c43835f5-9b6e-4b72-8c1d-0bcb32230845" (UID: "c43835f5-9b6e-4b72-8c1d-0bcb32230845"). InnerVolumeSpecName "kube-api-access-gzt9t". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.238000 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config" (OuterVolumeSpecName: "config") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.240823 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.246407 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.246481 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.249038 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59c87445-1528-4cab-9a2a-8d736c1a6c71" (UID: "59c87445-1528-4cab-9a2a-8d736c1a6c71"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.270537 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c43835f5-9b6e-4b72-8c1d-0bcb32230845" (UID: "c43835f5-9b6e-4b72-8c1d-0bcb32230845"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296397 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c43835f5-9b6e-4b72-8c1d-0bcb32230845-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296431 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzt9t\" (UniqueName: \"kubernetes.io/projected/c43835f5-9b6e-4b72-8c1d-0bcb32230845-kube-api-access-gzt9t\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296448 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g666h\" (UniqueName: \"kubernetes.io/projected/59c87445-1528-4cab-9a2a-8d736c1a6c71-kube-api-access-g666h\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296465 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296480 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296491 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296502 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.296512 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59c87445-1528-4cab-9a2a-8d736c1a6c71-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.431099 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-vcmh6" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.431243 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-vcmh6" event={"ID":"9174a689-38fd-478f-8c3e-07ab7e21a921","Type":"ContainerDied","Data":"49fc37dedeb9d41de6a45b462c37c6897c703a80b40d217ec187796af48c2362"} Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.431637 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49fc37dedeb9d41de6a45b462c37c6897c703a80b40d217ec187796af48c2362" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.434354 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5qfx" event={"ID":"c43835f5-9b6e-4b72-8c1d-0bcb32230845","Type":"ContainerDied","Data":"32c448398b7f7c739c9966fdcb1471f4ae5c5e20570a1e0b10c679b5debbd407"} Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.434530 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5qfx" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.439157 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.439189 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-648845bbf5-g7xjf" event={"ID":"59c87445-1528-4cab-9a2a-8d736c1a6c71","Type":"ContainerDied","Data":"85f7957b059b18a72ee551c44828ddcb26977b9335ab5d89274c91a699ecd802"} Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.440740 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.464129 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.469513 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.476447 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.494370 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5qfx"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.516880 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517386 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-log" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517408 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-log" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517422 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-httpd" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517429 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-httpd" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517442 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="registry-server" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517448 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="registry-server" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517459 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="extract-utilities" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517465 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="extract-utilities" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517481 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="init" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517487 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="init" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517499 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9174a689-38fd-478f-8c3e-07ab7e21a921" containerName="neutron-db-sync" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517505 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9174a689-38fd-478f-8c3e-07ab7e21a921" containerName="neutron-db-sync" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517513 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="extract-content" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517534 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="extract-content" Nov 25 05:55:09 crc kubenswrapper[4708]: E1125 05:55:09.517553 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="dnsmasq-dns" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517559 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="dnsmasq-dns" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517802 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="9174a689-38fd-478f-8c3e-07ab7e21a921" containerName="neutron-db-sync" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517818 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" containerName="dnsmasq-dns" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517830 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-log" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517845 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" containerName="glance-httpd" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.517856 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" containerName="registry-server" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.518963 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.520910 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.520976 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.533273 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.542297 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.547097 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-648845bbf5-g7xjf"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603070 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603150 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603335 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603412 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603574 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59g92\" (UniqueName: \"kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603634 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603670 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.603687 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705458 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705551 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705584 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705629 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59g92\" (UniqueName: \"kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705655 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705686 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705722 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.705943 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.706451 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.706709 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.717294 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.717665 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.718407 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.727222 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.731428 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59g92\" (UniqueName: \"kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.746572 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.748359 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.756367 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.770934 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.807671 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.807935 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.808147 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.808299 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nnxz\" (UniqueName: \"kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.808397 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.808463 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.833473 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910147 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910262 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910348 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910415 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910605 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nnxz\" (UniqueName: \"kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.910719 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.911271 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.911475 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.911691 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.911700 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.911762 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.924770 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nnxz\" (UniqueName: \"kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz\") pod \"dnsmasq-dns-6cb96b758c-xp5rs\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.968929 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.970703 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.975408 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-n6chh" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.975566 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.975835 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.982601 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:09 crc kubenswrapper[4708]: I1125 05:55:09.982747 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.012298 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.012370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.012454 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.012483 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.012581 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbg26\" (UniqueName: \"kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.116431 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.116513 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.116600 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbg26\" (UniqueName: \"kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.116671 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.116729 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.120986 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.121542 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.122533 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.122696 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.128399 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.134490 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbg26\" (UniqueName: \"kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26\") pod \"neutron-5d9dd8fd96-lxz2s\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.295353 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.901738 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04465ec9-6070-4ec3-89e0-c45b84056e60" path="/var/lib/kubelet/pods/04465ec9-6070-4ec3-89e0-c45b84056e60/volumes" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.902582 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59c87445-1528-4cab-9a2a-8d736c1a6c71" path="/var/lib/kubelet/pods/59c87445-1528-4cab-9a2a-8d736c1a6c71/volumes" Nov 25 05:55:10 crc kubenswrapper[4708]: I1125 05:55:10.903258 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c43835f5-9b6e-4b72-8c1d-0bcb32230845" path="/var/lib/kubelet/pods/c43835f5-9b6e-4b72-8c1d-0bcb32230845/volumes" Nov 25 05:55:11 crc kubenswrapper[4708]: I1125 05:55:11.939948 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-574d769d5-pkqlp"] Nov 25 05:55:11 crc kubenswrapper[4708]: I1125 05:55:11.943408 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:11 crc kubenswrapper[4708]: I1125 05:55:11.946076 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 05:55:11 crc kubenswrapper[4708]: I1125 05:55:11.946985 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 05:55:11 crc kubenswrapper[4708]: I1125 05:55:11.960702 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-574d769d5-pkqlp"] Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.063769 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-internal-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064262 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckjpr\" (UniqueName: \"kubernetes.io/projected/3abea71e-bc4c-4e36-a462-bd420d9b0090-kube-api-access-ckjpr\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064386 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-ovndb-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064592 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-httpd-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064652 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-combined-ca-bundle\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064839 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.064979 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-public-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166468 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-public-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166542 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-internal-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166626 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckjpr\" (UniqueName: \"kubernetes.io/projected/3abea71e-bc4c-4e36-a462-bd420d9b0090-kube-api-access-ckjpr\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166657 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-ovndb-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-httpd-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166707 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-combined-ca-bundle\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.166750 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.173696 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-ovndb-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.174588 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-public-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.174704 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-httpd-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.176039 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-internal-tls-certs\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.176373 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-combined-ca-bundle\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.179530 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3abea71e-bc4c-4e36-a462-bd420d9b0090-config\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.181924 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckjpr\" (UniqueName: \"kubernetes.io/projected/3abea71e-bc4c-4e36-a462-bd420d9b0090-kube-api-access-ckjpr\") pod \"neutron-574d769d5-pkqlp\" (UID: \"3abea71e-bc4c-4e36-a462-bd420d9b0090\") " pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:12 crc kubenswrapper[4708]: I1125 05:55:12.260201 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:17 crc kubenswrapper[4708]: I1125 05:55:17.265351 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:18 crc kubenswrapper[4708]: I1125 05:55:18.131209 4708 scope.go:117] "RemoveContainer" containerID="090cb4f191a00060e121de317f8c3f7cd94aa6006003adf807770b67f42e804a" Nov 25 05:55:18 crc kubenswrapper[4708]: I1125 05:55:18.442640 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8kfbn"] Nov 25 05:55:18 crc kubenswrapper[4708]: I1125 05:55:18.543295 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerStarted","Data":"f5f176a425bbde245b42076d84c58c4db477dc1210a8b9b0eb253e327eaea1c1"} Nov 25 05:55:18 crc kubenswrapper[4708]: I1125 05:55:18.605836 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:55:19 crc kubenswrapper[4708]: W1125 05:55:19.808714 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75e750a1_381d_4d38_8d4b_75ac9b3ea2d4.slice/crio-a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c WatchSource:0}: Error finding container a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c: Status 404 returned error can't find the container with id a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c Nov 25 05:55:19 crc kubenswrapper[4708]: I1125 05:55:19.828371 4708 scope.go:117] "RemoveContainer" containerID="de1fdaa727feb2782a730140b3db205f7dca6d9fabf5e9836dcc5e54e34ec73f" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.014107 4708 scope.go:117] "RemoveContainer" containerID="e9e50ec12ab64ad263f86fc144d31a0837b0c7be3015e00c6b534462aadea625" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.091616 4708 scope.go:117] "RemoveContainer" containerID="f38ca8faa43e41d183bca58a0da7dd333b2abd599cdad0cc03e37b376b19542d" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.126286 4708 scope.go:117] "RemoveContainer" containerID="e2f10046cbe7b3f4ca4d265a8d449c6d7eb8b25a7f96f3ed3c54f6078be495b9" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.346548 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-574d769d5-pkqlp"] Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.364120 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:20 crc kubenswrapper[4708]: W1125 05:55:20.368125 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9bc0dea1_eddd_4004_a450_e01c76354a12.slice/crio-69e4fcbb4daceb36eb691dba9de07039c2e98684823e8701e5803e1839e332bb WatchSource:0}: Error finding container 69e4fcbb4daceb36eb691dba9de07039c2e98684823e8701e5803e1839e332bb: Status 404 returned error can't find the container with id 69e4fcbb4daceb36eb691dba9de07039c2e98684823e8701e5803e1839e332bb Nov 25 05:55:20 crc kubenswrapper[4708]: W1125 05:55:20.379925 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3abea71e_bc4c_4e36_a462_bd420d9b0090.slice/crio-8028bf659e8b764f9ecbcad9d63436917903e80562d9482add16c94f496ec1b5 WatchSource:0}: Error finding container 8028bf659e8b764f9ecbcad9d63436917903e80562d9482add16c94f496ec1b5: Status 404 returned error can't find the container with id 8028bf659e8b764f9ecbcad9d63436917903e80562d9482add16c94f496ec1b5 Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.445092 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.577658 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lgvsg" event={"ID":"0f0e677f-bc35-4c02-8322-37d4cf7fc350","Type":"ContainerStarted","Data":"e7ff2d9e52ffd699b8820396f8c4afb7534c308c48c596fc2d0b5aa1b25b8abe"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.584382 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerStarted","Data":"1937efa3fda11f2cb34d4226959526d4124ad1333ab426883c904b112cf3b5ca"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.590405 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerStarted","Data":"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.599710 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerStarted","Data":"96cf11d3f1a806bfb19cad5acdfcce975fc5a1c5c4efd3d79c9aeb3dd9f48fdd"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.605282 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-lgvsg" podStartSLOduration=2.596985212 podStartE2EDuration="25.605267039s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="2025-11-25 05:54:56.801716888 +0000 UTC m=+838.210550274" lastFinishedPulling="2025-11-25 05:55:19.809998715 +0000 UTC m=+861.218832101" observedRunningTime="2025-11-25 05:55:20.596585779 +0000 UTC m=+862.005419165" watchObservedRunningTime="2025-11-25 05:55:20.605267039 +0000 UTC m=+862.014100426" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.638291 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerStarted","Data":"3401e4e693bc1228f1814338357cdfa36aeadc9a9e85c74f9460e3f2e6387a3d"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.641035 4708 generic.go:334] "Generic (PLEG): container finished" podID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerID="281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7" exitCode=0 Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.641133 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerDied","Data":"281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.663414 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2qhfv" podStartSLOduration=13.673386528 podStartE2EDuration="24.663389381s" podCreationTimestamp="2025-11-25 05:54:56 +0000 UTC" firstStartedPulling="2025-11-25 05:54:58.135097904 +0000 UTC m=+839.543931290" lastFinishedPulling="2025-11-25 05:55:09.125100757 +0000 UTC m=+850.533934143" observedRunningTime="2025-11-25 05:55:20.621090843 +0000 UTC m=+862.029924229" watchObservedRunningTime="2025-11-25 05:55:20.663389381 +0000 UTC m=+862.072222767" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.673691 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-574d769d5-pkqlp" event={"ID":"3abea71e-bc4c-4e36-a462-bd420d9b0090","Type":"ContainerStarted","Data":"8028bf659e8b764f9ecbcad9d63436917903e80562d9482add16c94f496ec1b5"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.689239 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8kfbn" event={"ID":"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4","Type":"ContainerStarted","Data":"e6e10051504f7f2c8358ecf0b9dc8076c4822fabbb4930beb6a6aacffc239809"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.689380 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8kfbn" event={"ID":"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4","Type":"ContainerStarted","Data":"a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.731886 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cqjrf" event={"ID":"5726473d-059a-4f7d-ac6b-4762c732f8ff","Type":"ContainerStarted","Data":"92d5c42cadaafa2a870118c54a28e34b934a292459ec7befc5120cf2cc463cf0"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.753993 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" event={"ID":"9bc0dea1-eddd-4004-a450-e01c76354a12","Type":"ContainerStarted","Data":"69e4fcbb4daceb36eb691dba9de07039c2e98684823e8701e5803e1839e332bb"} Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.767127 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cqjrf" podStartSLOduration=13.463038761 podStartE2EDuration="25.767109627s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="2025-11-25 05:54:56.819924247 +0000 UTC m=+838.228757634" lastFinishedPulling="2025-11-25 05:55:09.123995114 +0000 UTC m=+850.532828500" observedRunningTime="2025-11-25 05:55:20.763775644 +0000 UTC m=+862.172609020" watchObservedRunningTime="2025-11-25 05:55:20.767109627 +0000 UTC m=+862.175943013" Nov 25 05:55:20 crc kubenswrapper[4708]: I1125 05:55:20.787709 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8kfbn" podStartSLOduration=16.787690104 podStartE2EDuration="16.787690104s" podCreationTimestamp="2025-11-25 05:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:20.728715768 +0000 UTC m=+862.137549155" watchObservedRunningTime="2025-11-25 05:55:20.787690104 +0000 UTC m=+862.196523491" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.491109 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:55:21 crc kubenswrapper[4708]: W1125 05:55:21.495793 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod493b9467_e848_4aae_a5a5_5d699811e40e.slice/crio-ac1e9e7872f1b74c8133f0acfcabc2ef983467ddd07522256a4cc6634cdc00c8 WatchSource:0}: Error finding container ac1e9e7872f1b74c8133f0acfcabc2ef983467ddd07522256a4cc6634cdc00c8: Status 404 returned error can't find the container with id ac1e9e7872f1b74c8133f0acfcabc2ef983467ddd07522256a4cc6634cdc00c8 Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.784263 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-574d769d5-pkqlp" event={"ID":"3abea71e-bc4c-4e36-a462-bd420d9b0090","Type":"ContainerStarted","Data":"f4702568fb77128315603e4e9812aea4ef8c1075cce1ecd0bb3106b6c20abc47"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.784309 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-574d769d5-pkqlp" event={"ID":"3abea71e-bc4c-4e36-a462-bd420d9b0090","Type":"ContainerStarted","Data":"951c12cfd4c24b50798c5a5e3213fce8131aabb1dc458d94f2c6249524d7eb1f"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.784431 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.785943 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gnqch" event={"ID":"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f","Type":"ContainerStarted","Data":"37ab3898ffc69fdbb2cd13dd6500e287093eea52fa83a90c64f0f8f0323b8664"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.788592 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerStarted","Data":"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.788943 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerStarted","Data":"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.793761 4708 generic.go:334] "Generic (PLEG): container finished" podID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerID="915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f" exitCode=0 Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.793827 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerDied","Data":"915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.798292 4708 generic.go:334] "Generic (PLEG): container finished" podID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerID="e0d61cdd5f6b439ca1eef92b6ddb69108fb6ae4b4dd3dca39ee62d51bb46a30d" exitCode=0 Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.798507 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" event={"ID":"9bc0dea1-eddd-4004-a450-e01c76354a12","Type":"ContainerDied","Data":"e0d61cdd5f6b439ca1eef92b6ddb69108fb6ae4b4dd3dca39ee62d51bb46a30d"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.798571 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" event={"ID":"9bc0dea1-eddd-4004-a450-e01c76354a12","Type":"ContainerStarted","Data":"706c4378a7dac503ea6ce886b33cc513cbf033136c2efed12f3ec0bbe6946e94"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.798751 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.805572 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-574d769d5-pkqlp" podStartSLOduration=10.805560467 podStartE2EDuration="10.805560467s" podCreationTimestamp="2025-11-25 05:55:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:21.801984448 +0000 UTC m=+863.210817833" watchObservedRunningTime="2025-11-25 05:55:21.805560467 +0000 UTC m=+863.214393852" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.814493 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerStarted","Data":"6ea3ee053eea4bedd5ef86e724b5fde9e628b841551d9ee5f5a7ad7882df18d2"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.814562 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerStarted","Data":"463edbcae960768acc5fe0b24d2aceaf9a2fc9718b72705c4efea757e85b24b4"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.814746 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.818817 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerStarted","Data":"ac1e9e7872f1b74c8133f0acfcabc2ef983467ddd07522256a4cc6634cdc00c8"} Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.822404 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-gnqch" podStartSLOduration=3.355914263 podStartE2EDuration="26.822389074s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="2025-11-25 05:54:56.442385693 +0000 UTC m=+837.851219080" lastFinishedPulling="2025-11-25 05:55:19.908860505 +0000 UTC m=+861.317693891" observedRunningTime="2025-11-25 05:55:21.818928181 +0000 UTC m=+863.227761566" watchObservedRunningTime="2025-11-25 05:55:21.822389074 +0000 UTC m=+863.231222460" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.865725 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=15.865707063 podStartE2EDuration="15.865707063s" podCreationTimestamp="2025-11-25 05:55:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:21.853578274 +0000 UTC m=+863.262411660" watchObservedRunningTime="2025-11-25 05:55:21.865707063 +0000 UTC m=+863.274540448" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.881387 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" podStartSLOduration=12.881369190000001 podStartE2EDuration="12.88136919s" podCreationTimestamp="2025-11-25 05:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:21.871962873 +0000 UTC m=+863.280796259" watchObservedRunningTime="2025-11-25 05:55:21.88136919 +0000 UTC m=+863.290202577" Nov 25 05:55:21 crc kubenswrapper[4708]: I1125 05:55:21.902141 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5d9dd8fd96-lxz2s" podStartSLOduration=12.902128043 podStartE2EDuration="12.902128043s" podCreationTimestamp="2025-11-25 05:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:21.884608776 +0000 UTC m=+863.293442162" watchObservedRunningTime="2025-11-25 05:55:21.902128043 +0000 UTC m=+863.310961429" Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.837281 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerStarted","Data":"2d67c06df0ebc3b73444bb2776fef4baaa2f830f833c409a2743ec9d8f284268"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.840427 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerStarted","Data":"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.840482 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerStarted","Data":"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.844296 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerStarted","Data":"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.846915 4708 generic.go:334] "Generic (PLEG): container finished" podID="0f0e677f-bc35-4c02-8322-37d4cf7fc350" containerID="e7ff2d9e52ffd699b8820396f8c4afb7534c308c48c596fc2d0b5aa1b25b8abe" exitCode=0 Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.846973 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lgvsg" event={"ID":"0f0e677f-bc35-4c02-8322-37d4cf7fc350","Type":"ContainerDied","Data":"e7ff2d9e52ffd699b8820396f8c4afb7534c308c48c596fc2d0b5aa1b25b8abe"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.853315 4708 generic.go:334] "Generic (PLEG): container finished" podID="5726473d-059a-4f7d-ac6b-4762c732f8ff" containerID="92d5c42cadaafa2a870118c54a28e34b934a292459ec7befc5120cf2cc463cf0" exitCode=0 Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.853966 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cqjrf" event={"ID":"5726473d-059a-4f7d-ac6b-4762c732f8ff","Type":"ContainerDied","Data":"92d5c42cadaafa2a870118c54a28e34b934a292459ec7befc5120cf2cc463cf0"} Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.886222 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=13.886202012 podStartE2EDuration="13.886202012s" podCreationTimestamp="2025-11-25 05:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:22.867376882 +0000 UTC m=+864.276210288" watchObservedRunningTime="2025-11-25 05:55:22.886202012 +0000 UTC m=+864.295035398" Nov 25 05:55:22 crc kubenswrapper[4708]: I1125 05:55:22.925642 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-srn5s" podStartSLOduration=17.914268691 podStartE2EDuration="19.925622866s" podCreationTimestamp="2025-11-25 05:55:03 +0000 UTC" firstStartedPulling="2025-11-25 05:55:20.659877762 +0000 UTC m=+862.068711148" lastFinishedPulling="2025-11-25 05:55:22.671231947 +0000 UTC m=+864.080065323" observedRunningTime="2025-11-25 05:55:22.913260227 +0000 UTC m=+864.322093613" watchObservedRunningTime="2025-11-25 05:55:22.925622866 +0000 UTC m=+864.334456253" Nov 25 05:55:23 crc kubenswrapper[4708]: I1125 05:55:23.571299 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:23 crc kubenswrapper[4708]: I1125 05:55:23.571350 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.246922 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cqjrf" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.331575 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.337545 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data\") pod \"5726473d-059a-4f7d-ac6b-4762c732f8ff\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.337627 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts\") pod \"5726473d-059a-4f7d-ac6b-4762c732f8ff\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.337688 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grcbg\" (UniqueName: \"kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg\") pod \"5726473d-059a-4f7d-ac6b-4762c732f8ff\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.337718 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle\") pod \"5726473d-059a-4f7d-ac6b-4762c732f8ff\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.337856 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs\") pod \"5726473d-059a-4f7d-ac6b-4762c732f8ff\" (UID: \"5726473d-059a-4f7d-ac6b-4762c732f8ff\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.338449 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs" (OuterVolumeSpecName: "logs") pod "5726473d-059a-4f7d-ac6b-4762c732f8ff" (UID: "5726473d-059a-4f7d-ac6b-4762c732f8ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.338907 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5726473d-059a-4f7d-ac6b-4762c732f8ff-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.367713 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg" (OuterVolumeSpecName: "kube-api-access-grcbg") pod "5726473d-059a-4f7d-ac6b-4762c732f8ff" (UID: "5726473d-059a-4f7d-ac6b-4762c732f8ff"). InnerVolumeSpecName "kube-api-access-grcbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.369888 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts" (OuterVolumeSpecName: "scripts") pod "5726473d-059a-4f7d-ac6b-4762c732f8ff" (UID: "5726473d-059a-4f7d-ac6b-4762c732f8ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.417384 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5726473d-059a-4f7d-ac6b-4762c732f8ff" (UID: "5726473d-059a-4f7d-ac6b-4762c732f8ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.429800 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data" (OuterVolumeSpecName: "config-data") pod "5726473d-059a-4f7d-ac6b-4762c732f8ff" (UID: "5726473d-059a-4f7d-ac6b-4762c732f8ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.440125 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcxp7\" (UniqueName: \"kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7\") pod \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.440294 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data\") pod \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.440498 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle\") pod \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\" (UID: \"0f0e677f-bc35-4c02-8322-37d4cf7fc350\") " Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.441085 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.441104 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.441113 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grcbg\" (UniqueName: \"kubernetes.io/projected/5726473d-059a-4f7d-ac6b-4762c732f8ff-kube-api-access-grcbg\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.441123 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5726473d-059a-4f7d-ac6b-4762c732f8ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.444125 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7" (OuterVolumeSpecName: "kube-api-access-qcxp7") pod "0f0e677f-bc35-4c02-8322-37d4cf7fc350" (UID: "0f0e677f-bc35-4c02-8322-37d4cf7fc350"). InnerVolumeSpecName "kube-api-access-qcxp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.444423 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0f0e677f-bc35-4c02-8322-37d4cf7fc350" (UID: "0f0e677f-bc35-4c02-8322-37d4cf7fc350"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.461575 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f0e677f-bc35-4c02-8322-37d4cf7fc350" (UID: "0f0e677f-bc35-4c02-8322-37d4cf7fc350"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.542627 4708 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.542658 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f0e677f-bc35-4c02-8322-37d4cf7fc350-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.542668 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcxp7\" (UniqueName: \"kubernetes.io/projected/0f0e677f-bc35-4c02-8322-37d4cf7fc350-kube-api-access-qcxp7\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.626102 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-srn5s" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="registry-server" probeResult="failure" output=< Nov 25 05:55:24 crc kubenswrapper[4708]: timeout: failed to connect service ":50051" within 1s Nov 25 05:55:24 crc kubenswrapper[4708]: > Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.863975 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:24 crc kubenswrapper[4708]: E1125 05:55:24.864373 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f0e677f-bc35-4c02-8322-37d4cf7fc350" containerName="barbican-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.864395 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f0e677f-bc35-4c02-8322-37d4cf7fc350" containerName="barbican-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: E1125 05:55:24.864428 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5726473d-059a-4f7d-ac6b-4762c732f8ff" containerName="placement-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.864436 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="5726473d-059a-4f7d-ac6b-4762c732f8ff" containerName="placement-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.864644 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f0e677f-bc35-4c02-8322-37d4cf7fc350" containerName="barbican-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.864671 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="5726473d-059a-4f7d-ac6b-4762c732f8ff" containerName="placement-db-sync" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.865873 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.877411 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.885010 4708 generic.go:334] "Generic (PLEG): container finished" podID="e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" containerID="37ab3898ffc69fdbb2cd13dd6500e287093eea52fa83a90c64f0f8f0323b8664" exitCode=0 Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.885068 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gnqch" event={"ID":"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f","Type":"ContainerDied","Data":"37ab3898ffc69fdbb2cd13dd6500e287093eea52fa83a90c64f0f8f0323b8664"} Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.891106 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-lgvsg" event={"ID":"0f0e677f-bc35-4c02-8322-37d4cf7fc350","Type":"ContainerDied","Data":"bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a"} Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.891126 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-lgvsg" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.891136 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc9c9290ddc08431b3298ad17e997631af15bb3a7e3741006b70b715a02e0a3a" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.901775 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cqjrf" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.910256 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cqjrf" event={"ID":"5726473d-059a-4f7d-ac6b-4762c732f8ff","Type":"ContainerDied","Data":"63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b"} Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.910293 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63569d71ceddb9d60e4b909c8932e6e6d3e1f568e6939c9d145832d40b4a322b" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.915986 4708 generic.go:334] "Generic (PLEG): container finished" podID="75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" containerID="e6e10051504f7f2c8358ecf0b9dc8076c4822fabbb4930beb6a6aacffc239809" exitCode=0 Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.916046 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8kfbn" event={"ID":"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4","Type":"ContainerDied","Data":"e6e10051504f7f2c8358ecf0b9dc8076c4822fabbb4930beb6a6aacffc239809"} Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.951917 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.952014 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vn8bg\" (UniqueName: \"kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:24 crc kubenswrapper[4708]: I1125 05:55:24.952356 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.054797 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.054879 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vn8bg\" (UniqueName: \"kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.055033 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.055474 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.055536 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.074619 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vn8bg\" (UniqueName: \"kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg\") pod \"community-operators-fkw22\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.104590 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7b8ff6575b-8w7c4"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.114121 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.116288 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.117142 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-s5nlv" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.117270 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.117413 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.117573 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.148496 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b8ff6575b-8w7c4"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.193094 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.200595 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-8dcdb5d95-6tkvf"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.202624 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.216243 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-dhz8l" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.216458 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.216864 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.225100 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6db8477fbb-lpxf5"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.241969 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.259558 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263269 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-config-data\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263305 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263340 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kndp2\" (UniqueName: \"kubernetes.io/projected/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-kube-api-access-kndp2\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263417 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-combined-ca-bundle\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263441 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-internal-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263464 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-public-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263491 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-scripts\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263514 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-combined-ca-bundle\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263552 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-logs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263598 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data-custom\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263625 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92442175-4e83-4a5e-92ee-b75bfeca3f30-logs\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.263654 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfj49\" (UniqueName: \"kubernetes.io/projected/92442175-4e83-4a5e-92ee-b75bfeca3f30-kube-api-access-kfj49\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.342341 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8dcdb5d95-6tkvf"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.366073 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6db8477fbb-lpxf5"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.373894 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.373952 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wb95t\" (UniqueName: \"kubernetes.io/projected/06962f63-df39-4943-8021-1e0f4d337962-kube-api-access-wb95t\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374014 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06962f63-df39-4943-8021-1e0f4d337962-logs\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374042 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-combined-ca-bundle\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374066 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-internal-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374122 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-public-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374217 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-scripts\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374267 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-combined-ca-bundle\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374294 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-logs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374426 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-combined-ca-bundle\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374483 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data-custom\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374563 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92442175-4e83-4a5e-92ee-b75bfeca3f30-logs\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374623 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfj49\" (UniqueName: \"kubernetes.io/projected/92442175-4e83-4a5e-92ee-b75bfeca3f30-kube-api-access-kfj49\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374715 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-config-data\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374740 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374757 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data-custom\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.374809 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kndp2\" (UniqueName: \"kubernetes.io/projected/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-kube-api-access-kndp2\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.375545 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92442175-4e83-4a5e-92ee-b75bfeca3f30-logs\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.386738 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-logs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.388355 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.388976 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-internal-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.389662 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-combined-ca-bundle\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.390131 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-combined-ca-bundle\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.403765 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-scripts\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.407411 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-config-data\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.409022 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92442175-4e83-4a5e-92ee-b75bfeca3f30-config-data-custom\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.409080 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.409292 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="dnsmasq-dns" containerID="cri-o://706c4378a7dac503ea6ce886b33cc513cbf033136c2efed12f3ec0bbe6946e94" gracePeriod=10 Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.410135 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfj49\" (UniqueName: \"kubernetes.io/projected/92442175-4e83-4a5e-92ee-b75bfeca3f30-kube-api-access-kfj49\") pod \"barbican-worker-8dcdb5d95-6tkvf\" (UID: \"92442175-4e83-4a5e-92ee-b75bfeca3f30\") " pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.416503 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-public-tls-certs\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.444484 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kndp2\" (UniqueName: \"kubernetes.io/projected/17af25e9-9b9e-4175-b35c-18fe0b9e0c36-kube-api-access-kndp2\") pod \"placement-7b8ff6575b-8w7c4\" (UID: \"17af25e9-9b9e-4175-b35c-18fe0b9e0c36\") " pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.445147 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.451769 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.461598 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.461717 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.480194 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06962f63-df39-4943-8021-1e0f4d337962-logs\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.480401 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-combined-ca-bundle\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.480584 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data-custom\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.480672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.480712 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wb95t\" (UniqueName: \"kubernetes.io/projected/06962f63-df39-4943-8021-1e0f4d337962-kube-api-access-wb95t\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.481365 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/06962f63-df39-4943-8021-1e0f4d337962-logs\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.487853 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-combined-ca-bundle\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.496115 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data-custom\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.497163 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/06962f63-df39-4943-8021-1e0f4d337962-config-data\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.507746 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wb95t\" (UniqueName: \"kubernetes.io/projected/06962f63-df39-4943-8021-1e0f4d337962-kube-api-access-wb95t\") pod \"barbican-keystone-listener-6db8477fbb-lpxf5\" (UID: \"06962f63-df39-4943-8021-1e0f4d337962\") " pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.507796 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.509507 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.510819 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.565740 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.578897 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591697 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591777 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591802 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591833 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591865 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591902 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvnxh\" (UniqueName: \"kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591933 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bw4pg\" (UniqueName: \"kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.591975 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.592005 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.592063 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.592080 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.622885 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.693951 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.694001 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.694051 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.694898 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.694982 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695134 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695233 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695317 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvnxh\" (UniqueName: \"kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695400 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bw4pg\" (UniqueName: \"kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695537 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.695620 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.696474 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.696848 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.696988 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.697027 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.698317 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.698835 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.699598 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.701075 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.701645 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.713356 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvnxh\" (UniqueName: \"kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh\") pod \"barbican-api-78486ff96b-jvjsp\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.718250 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bw4pg\" (UniqueName: \"kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg\") pod \"dnsmasq-dns-864584b5c5-8v5rk\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.902096 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.909610 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.929365 4708 generic.go:334] "Generic (PLEG): container finished" podID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerID="706c4378a7dac503ea6ce886b33cc513cbf033136c2efed12f3ec0bbe6946e94" exitCode=0 Nov 25 05:55:25 crc kubenswrapper[4708]: I1125 05:55:25.929575 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" event={"ID":"9bc0dea1-eddd-4004-a450-e01c76354a12","Type":"ContainerDied","Data":"706c4378a7dac503ea6ce886b33cc513cbf033136c2efed12f3ec0bbe6946e94"} Nov 25 05:55:26 crc kubenswrapper[4708]: I1125 05:55:26.991481 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:26 crc kubenswrapper[4708]: I1125 05:55:26.991543 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.097205 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.097252 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.098371 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.140815 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.145552 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.520677 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646515 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646590 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mhr52\" (UniqueName: \"kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646651 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646721 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646753 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.646902 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys\") pod \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\" (UID: \"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4\") " Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.654725 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52" (OuterVolumeSpecName: "kube-api-access-mhr52") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "kube-api-access-mhr52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.655387 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.656635 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.673604 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts" (OuterVolumeSpecName: "scripts") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.707398 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data" (OuterVolumeSpecName: "config-data") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.710810 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" (UID: "75e750a1-381d-4d38-8d4b-75ac9b3ea2d4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.749944 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.749977 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mhr52\" (UniqueName: \"kubernetes.io/projected/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-kube-api-access-mhr52\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.749989 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.749999 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.750009 4708 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.750018 4708 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.963784 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8kfbn" event={"ID":"75e750a1-381d-4d38-8d4b-75ac9b3ea2d4","Type":"ContainerDied","Data":"a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c"} Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.964197 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9050b464993baecab3fd6a03f8f4ea9ee41f2fe8587e18d8babcbb49752b08c" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.963854 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8kfbn" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.965343 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 05:55:27 crc kubenswrapper[4708]: I1125 05:55:27.965386 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.026250 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.620036 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db6d644f4-2hdvr"] Nov 25 05:55:28 crc kubenswrapper[4708]: E1125 05:55:28.620984 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" containerName="keystone-bootstrap" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.621013 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" containerName="keystone-bootstrap" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.621212 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" containerName="keystone-bootstrap" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.622062 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.628272 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-sdz4c" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.628604 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.636991 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.637022 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.637008 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.637056 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.645317 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db6d644f4-2hdvr"] Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.771915 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-internal-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.771969 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-scripts\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772072 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2cj2\" (UniqueName: \"kubernetes.io/projected/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-kube-api-access-b2cj2\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772191 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-credential-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772227 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-fernet-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772302 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-combined-ca-bundle\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772414 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-public-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.772487 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-config-data\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.845573 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874360 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-combined-ca-bundle\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874414 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-public-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874479 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-config-data\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874591 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-internal-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874643 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-scripts\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874671 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2cj2\" (UniqueName: \"kubernetes.io/projected/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-kube-api-access-b2cj2\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874727 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-credential-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.874755 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-fernet-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.881057 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-credential-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.881198 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-public-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.881894 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-scripts\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.882185 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-fernet-keys\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.883494 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-config-data\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.886797 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-internal-tls-certs\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.889033 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-combined-ca-bundle\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.898475 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2cj2\" (UniqueName: \"kubernetes.io/projected/231fbeca-a7d9-4b28-b6ba-d1ed5ab61892-kube-api-access-b2cj2\") pod \"keystone-db6d644f4-2hdvr\" (UID: \"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892\") " pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:28 crc kubenswrapper[4708]: I1125 05:55:28.948320 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.101703 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-84bbd797c4-gvnft"] Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.104664 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.111670 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.111848 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.124562 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84bbd797c4-gvnft"] Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.181470 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-logs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.181540 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-public-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.181649 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lphbx\" (UniqueName: \"kubernetes.io/projected/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-kube-api-access-lphbx\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.181706 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.181901 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-internal-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.182022 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data-custom\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.182064 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-combined-ca-bundle\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.283907 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-internal-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284030 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data-custom\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284068 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-combined-ca-bundle\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284111 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-logs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284139 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-public-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284195 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lphbx\" (UniqueName: \"kubernetes.io/projected/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-kube-api-access-lphbx\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284242 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.284555 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-logs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.294489 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-combined-ca-bundle\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.294940 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-internal-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.298768 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data-custom\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.299487 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-config-data\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.302225 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lphbx\" (UniqueName: \"kubernetes.io/projected/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-kube-api-access-lphbx\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.310244 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/236c83bd-a2bb-4d2b-b43b-212abfb9eb1f-public-tls-certs\") pod \"barbican-api-84bbd797c4-gvnft\" (UID: \"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f\") " pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.437443 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.634659 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.637430 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.759824 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gnqch" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.764485 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.835423 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.835465 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909220 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909672 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909698 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909765 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909821 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909888 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909907 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nnxz\" (UniqueName: \"kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909954 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.909973 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.910092 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwvrl\" (UniqueName: \"kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.910112 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.910133 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0\") pod \"9bc0dea1-eddd-4004-a450-e01c76354a12\" (UID: \"9bc0dea1-eddd-4004-a450-e01c76354a12\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.910180 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle\") pod \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\" (UID: \"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f\") " Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.910917 4708 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.946822 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts" (OuterVolumeSpecName: "scripts") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.948369 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.950890 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl" (OuterVolumeSpecName: "kube-api-access-jwvrl") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "kube-api-access-jwvrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:29 crc kubenswrapper[4708]: I1125 05:55:29.951651 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz" (OuterVolumeSpecName: "kube-api-access-4nnxz") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "kube-api-access-4nnxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.012622 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwvrl\" (UniqueName: \"kubernetes.io/projected/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-kube-api-access-jwvrl\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.012651 4708 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.012660 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nnxz\" (UniqueName: \"kubernetes.io/projected/9bc0dea1-eddd-4004-a450-e01c76354a12-kube-api-access-4nnxz\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.012670 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.029106 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.029282 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" event={"ID":"9bc0dea1-eddd-4004-a450-e01c76354a12","Type":"ContainerDied","Data":"69e4fcbb4daceb36eb691dba9de07039c2e98684823e8701e5803e1839e332bb"} Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.029313 4708 scope.go:117] "RemoveContainer" containerID="706c4378a7dac503ea6ce886b33cc513cbf033136c2efed12f3ec0bbe6946e94" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.029941 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6cb96b758c-xp5rs" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.060189 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2qhfv" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="registry-server" containerID="cri-o://a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341" gracePeriod=2 Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.060771 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-gnqch" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.061314 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-gnqch" event={"ID":"e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f","Type":"ContainerDied","Data":"de1838a4044bddba3eb99f6a88a12f1627d70b4384f7b1d577fd41adbe7f54a3"} Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.061355 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de1838a4044bddba3eb99f6a88a12f1627d70b4384f7b1d577fd41adbe7f54a3" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.069877 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.114478 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.118260 4708 scope.go:117] "RemoveContainer" containerID="e0d61cdd5f6b439ca1eef92b6ddb69108fb6ae4b4dd3dca39ee62d51bb46a30d" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.122827 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.135921 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.160979 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data" (OuterVolumeSpecName: "config-data") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.161016 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" (UID: "e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.170167 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.173558 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.182250 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.198434 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config" (OuterVolumeSpecName: "config") pod "9bc0dea1-eddd-4004-a450-e01c76354a12" (UID: "9bc0dea1-eddd-4004-a450-e01c76354a12"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224621 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224648 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224658 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224668 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224678 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9bc0dea1-eddd-4004-a450-e01c76354a12-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.224687 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.362305 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.368260 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6cb96b758c-xp5rs"] Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.570636 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:30 crc kubenswrapper[4708]: W1125 05:55:30.572383 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66cc7b17_decb_414f_acc3_cd989e9246cc.slice/crio-a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41 WatchSource:0}: Error finding container a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41: Status 404 returned error can't find the container with id a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41 Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.880553 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db6d644f4-2hdvr"] Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.897326 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.925694 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" path="/var/lib/kubelet/pods/9bc0dea1-eddd-4004-a450-e01c76354a12/volumes" Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.936382 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.938653 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84bbd797c4-gvnft"] Nov 25 05:55:30 crc kubenswrapper[4708]: W1125 05:55:30.941350 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06962f63_df39_4943_8021_1e0f4d337962.slice/crio-ddc767c7f5a7241f79ccd9d751b5e1b0a993dec1dd3e5a5a0b72a573ae121bc3 WatchSource:0}: Error finding container ddc767c7f5a7241f79ccd9d751b5e1b0a993dec1dd3e5a5a0b72a573ae121bc3: Status 404 returned error can't find the container with id ddc767c7f5a7241f79ccd9d751b5e1b0a993dec1dd3e5a5a0b72a573ae121bc3 Nov 25 05:55:30 crc kubenswrapper[4708]: I1125 05:55:30.959219 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.016565 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-8dcdb5d95-6tkvf"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.035755 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6db8477fbb-lpxf5"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.046382 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxkm9\" (UniqueName: \"kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9\") pod \"d244e370-f811-42f5-9036-7ce6895289ab\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.046454 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities\") pod \"d244e370-f811-42f5-9036-7ce6895289ab\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.054204 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content\") pod \"d244e370-f811-42f5-9036-7ce6895289ab\" (UID: \"d244e370-f811-42f5-9036-7ce6895289ab\") " Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.054953 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities" (OuterVolumeSpecName: "utilities") pod "d244e370-f811-42f5-9036-7ce6895289ab" (UID: "d244e370-f811-42f5-9036-7ce6895289ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.096593 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7b8ff6575b-8w7c4"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.098200 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.101636 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9" (OuterVolumeSpecName: "kube-api-access-nxkm9") pod "d244e370-f811-42f5-9036-7ce6895289ab" (UID: "d244e370-f811-42f5-9036-7ce6895289ab"). InnerVolumeSpecName "kube-api-access-nxkm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.129632 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130604 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="dnsmasq-dns" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130630 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="dnsmasq-dns" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130649 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="init" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130660 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="init" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130677 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="extract-utilities" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130684 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="extract-utilities" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130722 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" containerName="cinder-db-sync" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130729 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" containerName="cinder-db-sync" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130754 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="extract-content" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130760 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="extract-content" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.130781 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="registry-server" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.130788 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="registry-server" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.131546 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" containerName="cinder-db-sync" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.131602 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="9bc0dea1-eddd-4004-a450-e01c76354a12" containerName="dnsmasq-dns" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.131620 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d244e370-f811-42f5-9036-7ce6895289ab" containerName="registry-server" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.133484 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.136984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84bbd797c4-gvnft" event={"ID":"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f","Type":"ContainerStarted","Data":"62d6474c8e081d52619ca75a33d0639f442de1d1408e2a22bfdf31f3619e63df"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.151300 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d244e370-f811-42f5-9036-7ce6895289ab" (UID: "d244e370-f811-42f5-9036-7ce6895289ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.153727 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db6d644f4-2hdvr" event={"ID":"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892","Type":"ContainerStarted","Data":"10df9b08322ea2323ec62a3b3ae3945540b97b4e306a36a88812f22eb1636cc2"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.158070 4708 generic.go:334] "Generic (PLEG): container finished" podID="d244e370-f811-42f5-9036-7ce6895289ab" containerID="a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341" exitCode=0 Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.158158 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerDied","Data":"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.158193 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qhfv" event={"ID":"d244e370-f811-42f5-9036-7ce6895289ab","Type":"ContainerDied","Data":"380fe082bb5cbddcc491b7db291801695e0632e94a032027780e80c78c1307a9"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.158224 4708 scope.go:117] "RemoveContainer" containerID="a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.158430 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qhfv" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.160160 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.160185 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.160543 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-np7px" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.160555 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.170984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" event={"ID":"06962f63-df39-4943-8021-1e0f4d337962","Type":"ContainerStarted","Data":"ddc767c7f5a7241f79ccd9d751b5e1b0a993dec1dd3e5a5a0b72a573ae121bc3"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.173653 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.182224 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerStarted","Data":"0b948510feaf019f35bf0bdc844fdb0d260336c825ec6ba77204a8b3870e8c07"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.184485 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" event={"ID":"92442175-4e83-4a5e-92ee-b75bfeca3f30","Type":"ContainerStarted","Data":"7cae6b8364c11dca1fb4149749ee9c3bb75accb5f43f7b31b78a69b20ed10b64"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.185312 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8ff6575b-8w7c4" event={"ID":"17af25e9-9b9e-4175-b35c-18fe0b9e0c36","Type":"ContainerStarted","Data":"1ee9fa5f833d4e83653c6015d17bcc58f6f3bfb131e4bcb892414578bfe38f73"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.190724 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" event={"ID":"66cc7b17-decb-414f-acc3-cd989e9246cc","Type":"ContainerStarted","Data":"6d729fdb92c1440837c1ccc1c0e5abb8fc64bf135e610f64c3d6a52763c6ba81"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.190753 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" event={"ID":"66cc7b17-decb-414f-acc3-cd989e9246cc","Type":"ContainerStarted","Data":"a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.194850 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.203555 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d244e370-f811-42f5-9036-7ce6895289ab-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.203584 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxkm9\" (UniqueName: \"kubernetes.io/projected/d244e370-f811-42f5-9036-7ce6895289ab-kube-api-access-nxkm9\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.205792 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerStarted","Data":"2d589cf3478a24f608edaf9700b43066e9730605849555f8bb193ca2ce4a9f61"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.217317 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.226587 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.228999 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.231932 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerStarted","Data":"5dc7cd5468e31e5bc20d088994c3d6bd3a038bd4c2eb72871f89b37c7317f817"} Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.232407 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.267891 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.269731 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.273218 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.279208 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.305902 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.305943 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cth4\" (UniqueName: \"kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.305985 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.306024 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.306043 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.306120 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.358735 4708 scope.go:117] "RemoveContainer" containerID="93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.391890 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.405126 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2qhfv"] Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.407570 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.407609 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.407809 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.407836 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cth4\" (UniqueName: \"kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.408240 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.408577 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.409618 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.410200 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.410500 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.411006 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.411067 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.411465 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.411501 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrqpj\" (UniqueName: \"kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412354 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412409 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412444 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412479 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412505 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412571 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lx9ns\" (UniqueName: \"kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.412686 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.424699 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cth4\" (UniqueName: \"kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.428249 4708 scope.go:117] "RemoveContainer" containerID="504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.428571 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.431741 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.432108 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.432413 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.454663 4708 scope.go:117] "RemoveContainer" containerID="a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.454967 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341\": container with ID starting with a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341 not found: ID does not exist" containerID="a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.454999 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341"} err="failed to get container status \"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341\": rpc error: code = NotFound desc = could not find container \"a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341\": container with ID starting with a9da18a6a225ed8779433283aebd1ef0cf585bf6f2b02880d1aa759128847341 not found: ID does not exist" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.455031 4708 scope.go:117] "RemoveContainer" containerID="93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.456346 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83\": container with ID starting with 93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83 not found: ID does not exist" containerID="93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.456370 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83"} err="failed to get container status \"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83\": rpc error: code = NotFound desc = could not find container \"93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83\": container with ID starting with 93a095c0e8a02ce9e41fb48cd04812c36b2beb4fd3e8fc9e44564cbcc7f24a83 not found: ID does not exist" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.456385 4708 scope.go:117] "RemoveContainer" containerID="504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.457153 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c\": container with ID starting with 504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c not found: ID does not exist" containerID="504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.457182 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c"} err="failed to get container status \"504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c\": rpc error: code = NotFound desc = could not find container \"504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c\": container with ID starting with 504597b619f32a0e7ffb63f745af7547a361ccc8e7c0157eb0a95647edd2ae6c not found: ID does not exist" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.489103 4708 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 25 05:55:31 crc kubenswrapper[4708]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/66cc7b17-decb-414f-acc3-cd989e9246cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 05:55:31 crc kubenswrapper[4708]: > podSandboxID="a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.489526 4708 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 05:55:31 crc kubenswrapper[4708]: container &Container{Name:dnsmasq-dns,Image:quay.rdoproject.org/podified-master-centos9/openstack-neutron-server:2cf1dc4bf18c6d57e990b3cd04e8ec78,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n7ch57ch5c5hcch589hf7h577h659h96h5c8h5b4h55fhbbh667h565h5bchcbh58dh7dh5bch586h56ch574h598h67dh5c8h56dh8bh574h564hbch7q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bw4pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-864584b5c5-8v5rk_openstack(66cc7b17-decb-414f-acc3-cd989e9246cc): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/66cc7b17-decb-414f-acc3-cd989e9246cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 05:55:31 crc kubenswrapper[4708]: > logger="UnhandledError" Nov 25 05:55:31 crc kubenswrapper[4708]: E1125 05:55:31.490859 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/66cc7b17-decb-414f-acc3-cd989e9246cc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" podUID="66cc7b17-decb-414f-acc3-cd989e9246cc" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513599 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513688 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513717 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513736 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrqpj\" (UniqueName: \"kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513764 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513784 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513800 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513815 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513843 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lx9ns\" (UniqueName: \"kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513863 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.513879 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.515308 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.515365 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.515808 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.516016 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.516168 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.516272 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.517355 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.517636 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.518036 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.519014 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.520788 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.528356 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrqpj\" (UniqueName: \"kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj\") pod \"cinder-api-0\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " pod="openstack/cinder-api-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.531694 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lx9ns\" (UniqueName: \"kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns\") pod \"dnsmasq-dns-6bf7f4798c-p7n82\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.682700 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.700113 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:31 crc kubenswrapper[4708]: I1125 05:55:31.711072 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.293146 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerStarted","Data":"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.293718 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerStarted","Data":"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.294981 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.295014 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.309802 4708 generic.go:334] "Generic (PLEG): container finished" podID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerID="0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05" exitCode=0 Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.309852 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerDied","Data":"0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.331064 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-78486ff96b-jvjsp" podStartSLOduration=7.331053021 podStartE2EDuration="7.331053021s" podCreationTimestamp="2025-11-25 05:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:32.317381085 +0000 UTC m=+873.726214471" watchObservedRunningTime="2025-11-25 05:55:32.331053021 +0000 UTC m=+873.739886407" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.363369 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84bbd797c4-gvnft" event={"ID":"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f","Type":"ContainerStarted","Data":"6465bca2acc424e27fc93391f4277c37e7c6e22a6443fe223e82012f3cbdb956"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.363419 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84bbd797c4-gvnft" event={"ID":"236c83bd-a2bb-4d2b-b43b-212abfb9eb1f","Type":"ContainerStarted","Data":"c73a2fe615c5a7d84c425d9f3353280b9706c34c402e94fc69e6ccf8505b6449"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.364350 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.364376 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.374643 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8ff6575b-8w7c4" event={"ID":"17af25e9-9b9e-4175-b35c-18fe0b9e0c36","Type":"ContainerStarted","Data":"0ce0b592255e99c7c09ed18d8d0711e5e7db6b0bf6c041501dce42fbeb776fa7"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.374673 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7b8ff6575b-8w7c4" event={"ID":"17af25e9-9b9e-4175-b35c-18fe0b9e0c36","Type":"ContainerStarted","Data":"a1bf5dc5d52f061d8497ac6d8538807ec102b04545471843f0cedc7da0180a24"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.375152 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.375175 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.399062 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-84bbd797c4-gvnft" podStartSLOduration=3.399050192 podStartE2EDuration="3.399050192s" podCreationTimestamp="2025-11-25 05:55:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:32.393722872 +0000 UTC m=+873.802556258" watchObservedRunningTime="2025-11-25 05:55:32.399050192 +0000 UTC m=+873.807883578" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.432462 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db6d644f4-2hdvr" event={"ID":"231fbeca-a7d9-4b28-b6ba-d1ed5ab61892","Type":"ContainerStarted","Data":"cdcc5a3b1fecdc485fc39a4e2c4ae651c9ec988b56ea0157e122039954b779e9"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.433125 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.446108 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7b8ff6575b-8w7c4" podStartSLOduration=7.446091358 podStartE2EDuration="7.446091358s" podCreationTimestamp="2025-11-25 05:55:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:32.433461826 +0000 UTC m=+873.842295213" watchObservedRunningTime="2025-11-25 05:55:32.446091358 +0000 UTC m=+873.854924745" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.456066 4708 generic.go:334] "Generic (PLEG): container finished" podID="66cc7b17-decb-414f-acc3-cd989e9246cc" containerID="6d729fdb92c1440837c1ccc1c0e5abb8fc64bf135e610f64c3d6a52763c6ba81" exitCode=0 Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.456135 4708 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.456813 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" event={"ID":"66cc7b17-decb-414f-acc3-cd989e9246cc","Type":"ContainerDied","Data":"6d729fdb92c1440837c1ccc1c0e5abb8fc64bf135e610f64c3d6a52763c6ba81"} Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.476064 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.523377 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db6d644f4-2hdvr" podStartSLOduration=4.523353771 podStartE2EDuration="4.523353771s" podCreationTimestamp="2025-11-25 05:55:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:32.464866524 +0000 UTC m=+873.873699909" watchObservedRunningTime="2025-11-25 05:55:32.523353771 +0000 UTC m=+873.932187157" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.551697 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.596837 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.744834 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.920177 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d244e370-f811-42f5-9036-7ce6895289ab" path="/var/lib/kubelet/pods/d244e370-f811-42f5-9036-7ce6895289ab/volumes" Nov 25 05:55:32 crc kubenswrapper[4708]: I1125 05:55:32.944367 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062067 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062208 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062318 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062358 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bw4pg\" (UniqueName: \"kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062392 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.062503 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0\") pod \"66cc7b17-decb-414f-acc3-cd989e9246cc\" (UID: \"66cc7b17-decb-414f-acc3-cd989e9246cc\") " Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.067722 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg" (OuterVolumeSpecName: "kube-api-access-bw4pg") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "kube-api-access-bw4pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.125128 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.136807 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.137269 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config" (OuterVolumeSpecName: "config") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.153882 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.155946 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "66cc7b17-decb-414f-acc3-cd989e9246cc" (UID: "66cc7b17-decb-414f-acc3-cd989e9246cc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165171 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165200 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bw4pg\" (UniqueName: \"kubernetes.io/projected/66cc7b17-decb-414f-acc3-cd989e9246cc-kube-api-access-bw4pg\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165212 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165222 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165232 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.165242 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66cc7b17-decb-414f-acc3-cd989e9246cc-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.423808 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.527776 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerStarted","Data":"45bccab35d641d8f292f514190b005001f6656bf0099033ec15b60b3d4419f9e"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.543154 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerStarted","Data":"f04ad94a434f3aabf72ade906e9de51118c04f6e0aafe5f93588f6373d28898d"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.545501 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" event={"ID":"66cc7b17-decb-414f-acc3-cd989e9246cc","Type":"ContainerDied","Data":"a686ef7b650c65e68b71d2c5bc981b3b9cf5afc8b40a81ea4ff0d3ad6431bf41"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.545562 4708 scope.go:117] "RemoveContainer" containerID="6d729fdb92c1440837c1ccc1c0e5abb8fc64bf135e610f64c3d6a52763c6ba81" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.545689 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-864584b5c5-8v5rk" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.555512 4708 generic.go:334] "Generic (PLEG): container finished" podID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerID="572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16" exitCode=0 Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.555593 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" event={"ID":"389c1d0d-ddc1-4113-9bdb-2142af81e18f","Type":"ContainerDied","Data":"572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.555613 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" event={"ID":"389c1d0d-ddc1-4113-9bdb-2142af81e18f","Type":"ContainerStarted","Data":"75c78060b6f5d959ef9847f38161736229b9e70c947641668131d5c8a9d5855d"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.564089 4708 generic.go:334] "Generic (PLEG): container finished" podID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerID="7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103" exitCode=0 Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.565177 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerDied","Data":"7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103"} Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.616956 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.634483 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-864584b5c5-8v5rk"] Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.672096 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:33 crc kubenswrapper[4708]: I1125 05:55:33.775858 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:34 crc kubenswrapper[4708]: I1125 05:55:34.053712 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:34 crc kubenswrapper[4708]: I1125 05:55:34.574531 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerStarted","Data":"114c9fefa4b9619fb5ad1bfed1bcac420a6664ca89cca71359e31eee55d0a120"} Nov 25 05:55:34 crc kubenswrapper[4708]: I1125 05:55:34.906328 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66cc7b17-decb-414f-acc3-cd989e9246cc" path="/var/lib/kubelet/pods/66cc7b17-decb-414f-acc3-cd989e9246cc/volumes" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.607858 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" event={"ID":"389c1d0d-ddc1-4113-9bdb-2142af81e18f","Type":"ContainerStarted","Data":"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.608309 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.630293 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" podStartSLOduration=4.630282655 podStartE2EDuration="4.630282655s" podCreationTimestamp="2025-11-25 05:55:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:35.624900442 +0000 UTC m=+877.033733828" watchObservedRunningTime="2025-11-25 05:55:35.630282655 +0000 UTC m=+877.039116042" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.634302 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerStarted","Data":"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.653982 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" event={"ID":"06962f63-df39-4943-8021-1e0f4d337962","Type":"ContainerStarted","Data":"e5c4dd35b9ae85778b95df89f8c8cf2f65d1f851681676c5a97d022522dace60"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.654011 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" event={"ID":"06962f63-df39-4943-8021-1e0f4d337962","Type":"ContainerStarted","Data":"92eaa8cb4c1e5615f9ab34d12526188a5b9beb56c23a682fc0b7aaef3737be7f"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.655727 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fkw22" podStartSLOduration=8.296003878 podStartE2EDuration="11.655713014s" podCreationTimestamp="2025-11-25 05:55:24 +0000 UTC" firstStartedPulling="2025-11-25 05:55:31.249957952 +0000 UTC m=+872.658791338" lastFinishedPulling="2025-11-25 05:55:34.609667089 +0000 UTC m=+876.018500474" observedRunningTime="2025-11-25 05:55:35.652842262 +0000 UTC m=+877.061675648" watchObservedRunningTime="2025-11-25 05:55:35.655713014 +0000 UTC m=+877.064546400" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.656796 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerStarted","Data":"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.658471 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" event={"ID":"92442175-4e83-4a5e-92ee-b75bfeca3f30","Type":"ContainerStarted","Data":"b670d404e377d81bdfe7e36c33d0413cb1009de4ab459737b28493808b31eca8"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.658499 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" event={"ID":"92442175-4e83-4a5e-92ee-b75bfeca3f30","Type":"ContainerStarted","Data":"d6a5d969d621f55b389726ebcd03b098b7370317fa7c001aac1a2b3324d788a3"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.669635 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerStarted","Data":"8de7aad14fa85aa9c4138dd80602381fc14940e65d0c1a408f6c2222ca6621e9"} Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.669760 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api-log" containerID="cri-o://114c9fefa4b9619fb5ad1bfed1bcac420a6664ca89cca71359e31eee55d0a120" gracePeriod=30 Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.670014 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.670078 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api" containerID="cri-o://8de7aad14fa85aa9c4138dd80602381fc14940e65d0c1a408f6c2222ca6621e9" gracePeriod=30 Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.671182 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6db8477fbb-lpxf5" podStartSLOduration=7.166936681 podStartE2EDuration="10.67117267s" podCreationTimestamp="2025-11-25 05:55:25 +0000 UTC" firstStartedPulling="2025-11-25 05:55:31.001714249 +0000 UTC m=+872.410547636" lastFinishedPulling="2025-11-25 05:55:34.505950239 +0000 UTC m=+875.914783625" observedRunningTime="2025-11-25 05:55:35.668511423 +0000 UTC m=+877.077344810" watchObservedRunningTime="2025-11-25 05:55:35.67117267 +0000 UTC m=+877.080006056" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.718257 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-8dcdb5d95-6tkvf" podStartSLOduration=7.216947131 podStartE2EDuration="10.718249763s" podCreationTimestamp="2025-11-25 05:55:25 +0000 UTC" firstStartedPulling="2025-11-25 05:55:31.006189653 +0000 UTC m=+872.415023039" lastFinishedPulling="2025-11-25 05:55:34.507492286 +0000 UTC m=+875.916325671" observedRunningTime="2025-11-25 05:55:35.696282572 +0000 UTC m=+877.105115959" watchObservedRunningTime="2025-11-25 05:55:35.718249763 +0000 UTC m=+877.127083149" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.750052 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.750027262 podStartE2EDuration="4.750027262s" podCreationTimestamp="2025-11-25 05:55:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:35.738567545 +0000 UTC m=+877.147400932" watchObservedRunningTime="2025-11-25 05:55:35.750027262 +0000 UTC m=+877.158860649" Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.847713 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:35 crc kubenswrapper[4708]: I1125 05:55:35.847975 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-srn5s" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="registry-server" containerID="cri-o://5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb" gracePeriod=2 Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.295782 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.365944 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities\") pod \"1989e5c5-4e78-4129-8562-779ea7ad5c47\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.366067 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dpxl\" (UniqueName: \"kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl\") pod \"1989e5c5-4e78-4129-8562-779ea7ad5c47\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.366174 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content\") pod \"1989e5c5-4e78-4129-8562-779ea7ad5c47\" (UID: \"1989e5c5-4e78-4129-8562-779ea7ad5c47\") " Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.367146 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities" (OuterVolumeSpecName: "utilities") pod "1989e5c5-4e78-4129-8562-779ea7ad5c47" (UID: "1989e5c5-4e78-4129-8562-779ea7ad5c47"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.372898 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl" (OuterVolumeSpecName: "kube-api-access-9dpxl") pod "1989e5c5-4e78-4129-8562-779ea7ad5c47" (UID: "1989e5c5-4e78-4129-8562-779ea7ad5c47"). InnerVolumeSpecName "kube-api-access-9dpxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.383928 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1989e5c5-4e78-4129-8562-779ea7ad5c47" (UID: "1989e5c5-4e78-4129-8562-779ea7ad5c47"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.468733 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.468766 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1989e5c5-4e78-4129-8562-779ea7ad5c47-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.468777 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dpxl\" (UniqueName: \"kubernetes.io/projected/1989e5c5-4e78-4129-8562-779ea7ad5c47-kube-api-access-9dpxl\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.683073 4708 generic.go:334] "Generic (PLEG): container finished" podID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerID="8de7aad14fa85aa9c4138dd80602381fc14940e65d0c1a408f6c2222ca6621e9" exitCode=0 Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.683336 4708 generic.go:334] "Generic (PLEG): container finished" podID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerID="114c9fefa4b9619fb5ad1bfed1bcac420a6664ca89cca71359e31eee55d0a120" exitCode=143 Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.683146 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerDied","Data":"8de7aad14fa85aa9c4138dd80602381fc14940e65d0c1a408f6c2222ca6621e9"} Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.683414 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerDied","Data":"114c9fefa4b9619fb5ad1bfed1bcac420a6664ca89cca71359e31eee55d0a120"} Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.686944 4708 generic.go:334] "Generic (PLEG): container finished" podID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerID="5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb" exitCode=0 Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.686982 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerDied","Data":"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb"} Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.687007 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-srn5s" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.687058 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-srn5s" event={"ID":"1989e5c5-4e78-4129-8562-779ea7ad5c47","Type":"ContainerDied","Data":"f5f176a425bbde245b42076d84c58c4db477dc1210a8b9b0eb253e327eaea1c1"} Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.687081 4708 scope.go:117] "RemoveContainer" containerID="5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.695552 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerStarted","Data":"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5"} Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.715009 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.733395985 podStartE2EDuration="5.714985124s" podCreationTimestamp="2025-11-25 05:55:31 +0000 UTC" firstStartedPulling="2025-11-25 05:55:32.507629667 +0000 UTC m=+873.916463053" lastFinishedPulling="2025-11-25 05:55:34.489218806 +0000 UTC m=+875.898052192" observedRunningTime="2025-11-25 05:55:36.7084408 +0000 UTC m=+878.117274187" watchObservedRunningTime="2025-11-25 05:55:36.714985124 +0000 UTC m=+878.123818509" Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.764259 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.773160 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-srn5s"] Nov 25 05:55:36 crc kubenswrapper[4708]: I1125 05:55:36.906321 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" path="/var/lib/kubelet/pods/1989e5c5-4e78-4129-8562-779ea7ad5c47/volumes" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.307005 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.314015 4708 scope.go:117] "RemoveContainer" containerID="915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.418730 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.473680 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.473968 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrqpj\" (UniqueName: \"kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474065 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474179 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474264 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474382 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id\") pod \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\" (UID: \"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea\") " Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.474985 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.476233 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs" (OuterVolumeSpecName: "logs") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.484242 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts" (OuterVolumeSpecName: "scripts") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.492918 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.511079 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj" (OuterVolumeSpecName: "kube-api-access-zrqpj") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "kube-api-access-zrqpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.522150 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.539415 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data" (OuterVolumeSpecName: "config-data") pod "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" (UID: "8ed834fe-cce2-4e08-97ed-bdd7d838a2ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578183 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578213 4708 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578223 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578234 4708 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578242 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578250 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrqpj\" (UniqueName: \"kubernetes.io/projected/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-kube-api-access-zrqpj\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.578261 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.598930 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.626714 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84bbd797c4-gvnft" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.698083 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.698374 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" containerID="cri-o://692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1" gracePeriod=30 Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.698507 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" containerID="cri-o://584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e" gracePeriod=30 Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.705846 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": EOF" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.705862 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": EOF" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.740115 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.740163 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"8ed834fe-cce2-4e08-97ed-bdd7d838a2ea","Type":"ContainerDied","Data":"f04ad94a434f3aabf72ade906e9de51118c04f6e0aafe5f93588f6373d28898d"} Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.772503 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.796643 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.817363 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.817913 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66cc7b17-decb-414f-acc3-cd989e9246cc" containerName="init" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.817936 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="66cc7b17-decb-414f-acc3-cd989e9246cc" containerName="init" Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.817950 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="extract-content" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.817958 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="extract-content" Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.817974 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="registry-server" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.817981 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="registry-server" Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.817995 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="extract-utilities" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818002 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="extract-utilities" Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.818023 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818029 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api" Nov 25 05:55:40 crc kubenswrapper[4708]: E1125 05:55:40.818039 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api-log" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818047 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api-log" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818248 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="66cc7b17-decb-414f-acc3-cd989e9246cc" containerName="init" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818265 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api-log" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818281 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1989e5c5-4e78-4129-8562-779ea7ad5c47" containerName="registry-server" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.818291 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" containerName="cinder-api" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.819426 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.825097 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.825410 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.825579 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.826175 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886567 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886669 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886716 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886791 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2nc7\" (UniqueName: \"kubernetes.io/projected/607ff629-5168-4560-a9bc-c4598573405b-kube-api-access-f2nc7\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886901 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-scripts\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.886977 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/607ff629-5168-4560-a9bc-c4598573405b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.887104 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data-custom\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.887161 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.887254 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607ff629-5168-4560-a9bc-c4598573405b-logs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.906149 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed834fe-cce2-4e08-97ed-bdd7d838a2ea" path="/var/lib/kubelet/pods/8ed834fe-cce2-4e08-97ed-bdd7d838a2ea/volumes" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.989909 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.989987 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990010 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2nc7\" (UniqueName: \"kubernetes.io/projected/607ff629-5168-4560-a9bc-c4598573405b-kube-api-access-f2nc7\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990031 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-scripts\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990064 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/607ff629-5168-4560-a9bc-c4598573405b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990556 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data-custom\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990625 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990721 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607ff629-5168-4560-a9bc-c4598573405b-logs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990803 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.990818 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/607ff629-5168-4560-a9bc-c4598573405b-etc-machine-id\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.993110 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/607ff629-5168-4560-a9bc-c4598573405b-logs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.993487 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-scripts\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.994959 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data-custom\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.994980 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.997144 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-public-tls-certs\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.997603 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-config-data\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:40 crc kubenswrapper[4708]: I1125 05:55:40.998497 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/607ff629-5168-4560-a9bc-c4598573405b-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.003955 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2nc7\" (UniqueName: \"kubernetes.io/projected/607ff629-5168-4560-a9bc-c4598573405b-kube-api-access-f2nc7\") pod \"cinder-api-0\" (UID: \"607ff629-5168-4560-a9bc-c4598573405b\") " pod="openstack/cinder-api-0" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.116901 4708 scope.go:117] "RemoveContainer" containerID="281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.138475 4708 scope.go:117] "RemoveContainer" containerID="5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb" Nov 25 05:55:41 crc kubenswrapper[4708]: E1125 05:55:41.138941 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb\": container with ID starting with 5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb not found: ID does not exist" containerID="5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.138996 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb"} err="failed to get container status \"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb\": rpc error: code = NotFound desc = could not find container \"5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb\": container with ID starting with 5f688d4a138bdc04cc12486343e504f765d204dcc1f7123fcd56207fdc11effb not found: ID does not exist" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.139025 4708 scope.go:117] "RemoveContainer" containerID="915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.139297 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 05:55:41 crc kubenswrapper[4708]: E1125 05:55:41.139898 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f\": container with ID starting with 915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f not found: ID does not exist" containerID="915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.139930 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f"} err="failed to get container status \"915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f\": rpc error: code = NotFound desc = could not find container \"915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f\": container with ID starting with 915ef5fba4176a11d2d23b8c71273bf0d3956b336d83b0acfa33db561c7c863f not found: ID does not exist" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.139951 4708 scope.go:117] "RemoveContainer" containerID="281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7" Nov 25 05:55:41 crc kubenswrapper[4708]: E1125 05:55:41.140493 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7\": container with ID starting with 281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7 not found: ID does not exist" containerID="281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.140551 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7"} err="failed to get container status \"281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7\": rpc error: code = NotFound desc = could not find container \"281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7\": container with ID starting with 281f3a6cbb4e0a714fc1bfdf01b9de4fe880cc86c6fd825c4472a131d05df4d7 not found: ID does not exist" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.140582 4708 scope.go:117] "RemoveContainer" containerID="8de7aad14fa85aa9c4138dd80602381fc14940e65d0c1a408f6c2222ca6621e9" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.176231 4708 scope.go:117] "RemoveContainer" containerID="114c9fefa4b9619fb5ad1bfed1bcac420a6664ca89cca71359e31eee55d0a120" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.673014 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.683776 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.702654 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.799857 4708 generic.go:334] "Generic (PLEG): container finished" podID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerID="692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1" exitCode=143 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.799933 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerDied","Data":"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1"} Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.826260 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.826486 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="dnsmasq-dns" containerID="cri-o://ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd" gracePeriod=10 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.851266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerStarted","Data":"2ed5d9acde6bdb404e2f5858995dcc4cefead4a8cd61ffb226f0b5c6b55c41d8"} Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.851479 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-central-agent" containerID="cri-o://1937efa3fda11f2cb34d4226959526d4124ad1333ab426883c904b112cf3b5ca" gracePeriod=30 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.851723 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.851990 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="proxy-httpd" containerID="cri-o://2ed5d9acde6bdb404e2f5858995dcc4cefead4a8cd61ffb226f0b5c6b55c41d8" gracePeriod=30 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.852045 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="sg-core" containerID="cri-o://0b948510feaf019f35bf0bdc844fdb0d260336c825ec6ba77204a8b3870e8c07" gracePeriod=30 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.852091 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-notification-agent" containerID="cri-o://2d67c06df0ebc3b73444bb2776fef4baaa2f830f833c409a2743ec9d8f284268" gracePeriod=30 Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.855755 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"607ff629-5168-4560-a9bc-c4598573405b","Type":"ContainerStarted","Data":"488d97f48b69903cafa58fd04976aab9d6a246c5c3e417eddcd0ed3957c9d67f"} Nov 25 05:55:41 crc kubenswrapper[4708]: I1125 05:55:41.895555 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.86388001 podStartE2EDuration="46.895536256s" podCreationTimestamp="2025-11-25 05:54:55 +0000 UTC" firstStartedPulling="2025-11-25 05:54:56.108047378 +0000 UTC m=+837.516880764" lastFinishedPulling="2025-11-25 05:55:41.139703624 +0000 UTC m=+882.548537010" observedRunningTime="2025-11-25 05:55:41.884904689 +0000 UTC m=+883.293738076" watchObservedRunningTime="2025-11-25 05:55:41.895536256 +0000 UTC m=+883.304369642" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.202877 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.250085 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.292606 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-574d769d5-pkqlp" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.410148 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.410800 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9dd8fd96-lxz2s" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-api" containerID="cri-o://463edbcae960768acc5fe0b24d2aceaf9a2fc9718b72705c4efea757e85b24b4" gracePeriod=30 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.411009 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5d9dd8fd96-lxz2s" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-httpd" containerID="cri-o://6ea3ee053eea4bedd5ef86e724b5fde9e628b841551d9ee5f5a7ad7882df18d2" gracePeriod=30 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.430408 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544409 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rhx7\" (UniqueName: \"kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544469 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544498 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544646 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544745 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.544790 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb\") pod \"454358fb-f82d-4c19-b6e3-5f1176829d6e\" (UID: \"454358fb-f82d-4c19-b6e3-5f1176829d6e\") " Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.565290 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7" (OuterVolumeSpecName: "kube-api-access-4rhx7") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "kube-api-access-4rhx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.623201 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config" (OuterVolumeSpecName: "config") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.636415 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.646699 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rhx7\" (UniqueName: \"kubernetes.io/projected/454358fb-f82d-4c19-b6e3-5f1176829d6e-kube-api-access-4rhx7\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.646726 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.646738 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.647413 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.654500 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.659492 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "454358fb-f82d-4c19-b6e3-5f1176829d6e" (UID: "454358fb-f82d-4c19-b6e3-5f1176829d6e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.748128 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.748160 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.748171 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/454358fb-f82d-4c19-b6e3-5f1176829d6e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.877385 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"607ff629-5168-4560-a9bc-c4598573405b","Type":"ContainerStarted","Data":"18c9a54f2915c8854896cf201cc790a4b84a9141cb8d98f234d2240bcf77f54c"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.880357 4708 generic.go:334] "Generic (PLEG): container finished" podID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerID="ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd" exitCode=0 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.880430 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" event={"ID":"454358fb-f82d-4c19-b6e3-5f1176829d6e","Type":"ContainerDied","Data":"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.880440 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.880462 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-69d79494c5-xf4kn" event={"ID":"454358fb-f82d-4c19-b6e3-5f1176829d6e","Type":"ContainerDied","Data":"02efa37391ee85e114fa802e9c2fd2b53ccec4638929bbe55422a1011bc86a0e"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.880483 4708 scope.go:117] "RemoveContainer" containerID="ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889218 4708 generic.go:334] "Generic (PLEG): container finished" podID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerID="2ed5d9acde6bdb404e2f5858995dcc4cefead4a8cd61ffb226f0b5c6b55c41d8" exitCode=0 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889240 4708 generic.go:334] "Generic (PLEG): container finished" podID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerID="0b948510feaf019f35bf0bdc844fdb0d260336c825ec6ba77204a8b3870e8c07" exitCode=2 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889248 4708 generic.go:334] "Generic (PLEG): container finished" podID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerID="1937efa3fda11f2cb34d4226959526d4124ad1333ab426883c904b112cf3b5ca" exitCode=0 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889287 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerDied","Data":"2ed5d9acde6bdb404e2f5858995dcc4cefead4a8cd61ffb226f0b5c6b55c41d8"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889307 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerDied","Data":"0b948510feaf019f35bf0bdc844fdb0d260336c825ec6ba77204a8b3870e8c07"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.889318 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerDied","Data":"1937efa3fda11f2cb34d4226959526d4124ad1333ab426883c904b112cf3b5ca"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.907295 4708 generic.go:334] "Generic (PLEG): container finished" podID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerID="6ea3ee053eea4bedd5ef86e724b5fde9e628b841551d9ee5f5a7ad7882df18d2" exitCode=0 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.907573 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="cinder-scheduler" containerID="cri-o://6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9" gracePeriod=30 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.908147 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="probe" containerID="cri-o://a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5" gracePeriod=30 Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.914328 4708 scope.go:117] "RemoveContainer" containerID="72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.917346 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.917396 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerDied","Data":"6ea3ee053eea4bedd5ef86e724b5fde9e628b841551d9ee5f5a7ad7882df18d2"} Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.923123 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-69d79494c5-xf4kn"] Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.936402 4708 scope.go:117] "RemoveContainer" containerID="ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd" Nov 25 05:55:42 crc kubenswrapper[4708]: E1125 05:55:42.938050 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd\": container with ID starting with ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd not found: ID does not exist" containerID="ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.938109 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd"} err="failed to get container status \"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd\": rpc error: code = NotFound desc = could not find container \"ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd\": container with ID starting with ee3db0919023eebe517e7aaff6376a3141029f4b11a1dbdb0aabcd61afd840fd not found: ID does not exist" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.938141 4708 scope.go:117] "RemoveContainer" containerID="72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca" Nov 25 05:55:42 crc kubenswrapper[4708]: E1125 05:55:42.938506 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca\": container with ID starting with 72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca not found: ID does not exist" containerID="72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca" Nov 25 05:55:42 crc kubenswrapper[4708]: I1125 05:55:42.938586 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca"} err="failed to get container status \"72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca\": rpc error: code = NotFound desc = could not find container \"72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca\": container with ID starting with 72385d398fe97975e9767ba1035de6558f4198844e47099e19e85e3a1ccf6fca not found: ID does not exist" Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.923376 4708 generic.go:334] "Generic (PLEG): container finished" podID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerID="2d67c06df0ebc3b73444bb2776fef4baaa2f830f833c409a2743ec9d8f284268" exitCode=0 Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.923464 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerDied","Data":"2d67c06df0ebc3b73444bb2776fef4baaa2f830f833c409a2743ec9d8f284268"} Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.925884 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"607ff629-5168-4560-a9bc-c4598573405b","Type":"ContainerStarted","Data":"f3c936601af7061b013566e2c462fe590481df207b6b1becd8fdeee885af2d4c"} Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.926014 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.927665 4708 generic.go:334] "Generic (PLEG): container finished" podID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerID="a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5" exitCode=0 Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.927735 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerDied","Data":"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5"} Nov 25 05:55:43 crc kubenswrapper[4708]: I1125 05:55:43.949171 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.9491207409999998 podStartE2EDuration="3.949120741s" podCreationTimestamp="2025-11-25 05:55:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:43.943901646 +0000 UTC m=+885.352735021" watchObservedRunningTime="2025-11-25 05:55:43.949120741 +0000 UTC m=+885.357954128" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.032905 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173244 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbk69\" (UniqueName: \"kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173327 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173381 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173457 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173485 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173625 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173730 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle\") pod \"d07de339-30ce-4edc-86f5-2f2dcebc417e\" (UID: \"d07de339-30ce-4edc-86f5-2f2dcebc417e\") " Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.173882 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.174563 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.174996 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.175019 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d07de339-30ce-4edc-86f5-2f2dcebc417e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.178873 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts" (OuterVolumeSpecName: "scripts") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.178922 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69" (OuterVolumeSpecName: "kube-api-access-sbk69") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "kube-api-access-sbk69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.196441 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.232640 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.249031 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data" (OuterVolumeSpecName: "config-data") pod "d07de339-30ce-4edc-86f5-2f2dcebc417e" (UID: "d07de339-30ce-4edc-86f5-2f2dcebc417e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.276637 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.276665 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbk69\" (UniqueName: \"kubernetes.io/projected/d07de339-30ce-4edc-86f5-2f2dcebc417e-kube-api-access-sbk69\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.276679 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.276690 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.276699 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d07de339-30ce-4edc-86f5-2f2dcebc417e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.357111 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.357192 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.905127 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" path="/var/lib/kubelet/pods/454358fb-f82d-4c19-b6e3-5f1176829d6e/volumes" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.942044 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d07de339-30ce-4edc-86f5-2f2dcebc417e","Type":"ContainerDied","Data":"5c1b127e10d58661b19d328167cfc361999745ee6dad5810a50d3569e9d887b8"} Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.942092 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.942760 4708 scope.go:117] "RemoveContainer" containerID="2ed5d9acde6bdb404e2f5858995dcc4cefead4a8cd61ffb226f0b5c6b55c41d8" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.972892 4708 scope.go:117] "RemoveContainer" containerID="0b948510feaf019f35bf0bdc844fdb0d260336c825ec6ba77204a8b3870e8c07" Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.980964 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.994455 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:55:44 crc kubenswrapper[4708]: I1125 05:55:44.994622 4708 scope.go:117] "RemoveContainer" containerID="2d67c06df0ebc3b73444bb2776fef4baaa2f830f833c409a2743ec9d8f284268" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011302 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011811 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-notification-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011831 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-notification-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011846 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="sg-core" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011853 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="sg-core" Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011865 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="dnsmasq-dns" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011872 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="dnsmasq-dns" Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011888 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-central-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011893 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-central-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011925 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="proxy-httpd" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011931 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="proxy-httpd" Nov 25 05:55:45 crc kubenswrapper[4708]: E1125 05:55:45.011938 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="init" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.011943 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="init" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.012098 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="proxy-httpd" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.012116 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="454358fb-f82d-4c19-b6e3-5f1176829d6e" containerName="dnsmasq-dns" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.012126 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="sg-core" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.012136 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-notification-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.012147 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" containerName="ceilometer-central-agent" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.013743 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.017029 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.020045 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.022626 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.023681 4708 scope.go:117] "RemoveContainer" containerID="1937efa3fda11f2cb34d4226959526d4124ad1333ab426883c904b112cf3b5ca" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.097918 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098137 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098165 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7rrr\" (UniqueName: \"kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098262 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098330 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098444 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.098579 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.194854 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.194891 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.199639 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.199686 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.199752 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200233 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200296 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200563 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200623 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200707 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.200726 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7rrr\" (UniqueName: \"kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.204441 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.205010 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.208071 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.218945 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.222171 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7rrr\" (UniqueName: \"kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr\") pod \"ceilometer-0\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.236374 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.334664 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.794739 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.952952 4708 generic.go:334] "Generic (PLEG): container finished" podID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerID="463edbcae960768acc5fe0b24d2aceaf9a2fc9718b72705c4efea757e85b24b4" exitCode=0 Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.953028 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerDied","Data":"463edbcae960768acc5fe0b24d2aceaf9a2fc9718b72705c4efea757e85b24b4"} Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.954049 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerStarted","Data":"4ca2eba6fe9e1a67e738f51be3bd8a0b24eb7202f80f1866914844b5b6bf1158"} Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.991379 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.992677 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 05:55:45 crc kubenswrapper[4708]: I1125 05:55:45.992687 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.031998 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.102055 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": read tcp 10.217.0.2:37790->10.217.0.153:9311: read: connection reset by peer" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.102098 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-78486ff96b-jvjsp" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.153:9311/healthcheck\": read tcp 10.217.0.2:37782->10.217.0.153:9311: read: connection reset by peer" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.239781 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.299254 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425078 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbg26\" (UniqueName: \"kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26\") pod \"a3070821-6f6c-40fa-9755-e46c6651ee92\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425187 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config\") pod \"a3070821-6f6c-40fa-9755-e46c6651ee92\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425219 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cth4\" (UniqueName: \"kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425252 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle\") pod \"a3070821-6f6c-40fa-9755-e46c6651ee92\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425299 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425344 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425501 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config\") pod \"a3070821-6f6c-40fa-9755-e46c6651ee92\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425602 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs\") pod \"a3070821-6f6c-40fa-9755-e46c6651ee92\" (UID: \"a3070821-6f6c-40fa-9755-e46c6651ee92\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425698 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425735 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.425885 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom\") pod \"e208d102-59a4-4a0e-8041-84a4df5e4433\" (UID: \"e208d102-59a4-4a0e-8041-84a4df5e4433\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.426323 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.427050 4708 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e208d102-59a4-4a0e-8041-84a4df5e4433-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.432895 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts" (OuterVolumeSpecName: "scripts") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.432908 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26" (OuterVolumeSpecName: "kube-api-access-vbg26") pod "a3070821-6f6c-40fa-9755-e46c6651ee92" (UID: "a3070821-6f6c-40fa-9755-e46c6651ee92"). InnerVolumeSpecName "kube-api-access-vbg26". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.433548 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.436099 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4" (OuterVolumeSpecName: "kube-api-access-4cth4") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "kube-api-access-4cth4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.442966 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a3070821-6f6c-40fa-9755-e46c6651ee92" (UID: "a3070821-6f6c-40fa-9755-e46c6651ee92"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.473353 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config" (OuterVolumeSpecName: "config") pod "a3070821-6f6c-40fa-9755-e46c6651ee92" (UID: "a3070821-6f6c-40fa-9755-e46c6651ee92"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.477213 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a3070821-6f6c-40fa-9755-e46c6651ee92" (UID: "a3070821-6f6c-40fa-9755-e46c6651ee92"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.479018 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.494633 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a3070821-6f6c-40fa-9755-e46c6651ee92" (UID: "a3070821-6f6c-40fa-9755-e46c6651ee92"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.517108 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data" (OuterVolumeSpecName: "config-data") pod "e208d102-59a4-4a0e-8041-84a4df5e4433" (UID: "e208d102-59a4-4a0e-8041-84a4df5e4433"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528549 4708 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528645 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbg26\" (UniqueName: \"kubernetes.io/projected/a3070821-6f6c-40fa-9755-e46c6651ee92-kube-api-access-vbg26\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528707 4708 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528771 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cth4\" (UniqueName: \"kubernetes.io/projected/e208d102-59a4-4a0e-8041-84a4df5e4433-kube-api-access-4cth4\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528823 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528871 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528925 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.528974 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.529023 4708 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a3070821-6f6c-40fa-9755-e46c6651ee92-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.529073 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e208d102-59a4-4a0e-8041-84a4df5e4433-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.655170 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.834726 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lvnxh\" (UniqueName: \"kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh\") pod \"6154c760-6930-4ceb-99d0-2aee2bf196ce\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.834852 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle\") pod \"6154c760-6930-4ceb-99d0-2aee2bf196ce\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.834942 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom\") pod \"6154c760-6930-4ceb-99d0-2aee2bf196ce\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.835087 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs\") pod \"6154c760-6930-4ceb-99d0-2aee2bf196ce\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.835197 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data\") pod \"6154c760-6930-4ceb-99d0-2aee2bf196ce\" (UID: \"6154c760-6930-4ceb-99d0-2aee2bf196ce\") " Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.835755 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs" (OuterVolumeSpecName: "logs") pod "6154c760-6930-4ceb-99d0-2aee2bf196ce" (UID: "6154c760-6930-4ceb-99d0-2aee2bf196ce"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.835995 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6154c760-6930-4ceb-99d0-2aee2bf196ce-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.839920 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6154c760-6930-4ceb-99d0-2aee2bf196ce" (UID: "6154c760-6930-4ceb-99d0-2aee2bf196ce"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.840308 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh" (OuterVolumeSpecName: "kube-api-access-lvnxh") pod "6154c760-6930-4ceb-99d0-2aee2bf196ce" (UID: "6154c760-6930-4ceb-99d0-2aee2bf196ce"). InnerVolumeSpecName "kube-api-access-lvnxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.858142 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6154c760-6930-4ceb-99d0-2aee2bf196ce" (UID: "6154c760-6930-4ceb-99d0-2aee2bf196ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.874260 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data" (OuterVolumeSpecName: "config-data") pod "6154c760-6930-4ceb-99d0-2aee2bf196ce" (UID: "6154c760-6930-4ceb-99d0-2aee2bf196ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.902058 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d07de339-30ce-4edc-86f5-2f2dcebc417e" path="/var/lib/kubelet/pods/d07de339-30ce-4edc-86f5-2f2dcebc417e/volumes" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.937641 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.937671 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lvnxh\" (UniqueName: \"kubernetes.io/projected/6154c760-6930-4ceb-99d0-2aee2bf196ce-kube-api-access-lvnxh\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.937683 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.937693 4708 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6154c760-6930-4ceb-99d0-2aee2bf196ce-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.966093 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerStarted","Data":"11e7c10b8784262fe65ec4e3c68b654129ff238dffe7b4525c2fc1ca2d0277e3"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.967774 4708 generic.go:334] "Generic (PLEG): container finished" podID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerID="584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e" exitCode=0 Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.967831 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-78486ff96b-jvjsp" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.967851 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerDied","Data":"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.968118 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-78486ff96b-jvjsp" event={"ID":"6154c760-6930-4ceb-99d0-2aee2bf196ce","Type":"ContainerDied","Data":"2d589cf3478a24f608edaf9700b43066e9730605849555f8bb193ca2ce4a9f61"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.968154 4708 scope.go:117] "RemoveContainer" containerID="584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.970702 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5d9dd8fd96-lxz2s" event={"ID":"a3070821-6f6c-40fa-9755-e46c6651ee92","Type":"ContainerDied","Data":"96cf11d3f1a806bfb19cad5acdfcce975fc5a1c5c4efd3d79c9aeb3dd9f48fdd"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.970859 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5d9dd8fd96-lxz2s" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.973773 4708 generic.go:334] "Generic (PLEG): container finished" podID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerID="6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9" exitCode=0 Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.974386 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.974809 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerDied","Data":"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.974839 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e208d102-59a4-4a0e-8041-84a4df5e4433","Type":"ContainerDied","Data":"45bccab35d641d8f292f514190b005001f6656bf0099033ec15b60b3d4419f9e"} Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.995280 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:46 crc kubenswrapper[4708]: I1125 05:55:46.996275 4708 scope.go:117] "RemoveContainer" containerID="692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.005743 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5d9dd8fd96-lxz2s"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.011175 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.023720 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-78486ff96b-jvjsp"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.027001 4708 scope.go:117] "RemoveContainer" containerID="584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.028642 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e\": container with ID starting with 584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e not found: ID does not exist" containerID="584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.028733 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e"} err="failed to get container status \"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e\": rpc error: code = NotFound desc = could not find container \"584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e\": container with ID starting with 584928811edbe15d512df0a6d2a0899514c96d8738cb57a77956a67459c7a81e not found: ID does not exist" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.028810 4708 scope.go:117] "RemoveContainer" containerID="692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.029800 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1\": container with ID starting with 692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1 not found: ID does not exist" containerID="692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.029850 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1"} err="failed to get container status \"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1\": rpc error: code = NotFound desc = could not find container \"692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1\": container with ID starting with 692338439f05d0706318b6810d534ba3c9e263f482c25bc332c01667308115c1 not found: ID does not exist" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.029884 4708 scope.go:117] "RemoveContainer" containerID="6ea3ee053eea4bedd5ef86e724b5fde9e628b841551d9ee5f5a7ad7882df18d2" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.033479 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.043601 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.048561 4708 scope.go:117] "RemoveContainer" containerID="463edbcae960768acc5fe0b24d2aceaf9a2fc9718b72705c4efea757e85b24b4" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.049301 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050016 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050031 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050051 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="cinder-scheduler" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050059 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="cinder-scheduler" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050083 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-api" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050089 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-api" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050103 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050109 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050119 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="probe" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050133 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="probe" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.050152 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-httpd" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050158 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-httpd" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050319 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-api" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050333 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" containerName="neutron-httpd" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050341 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api-log" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050349 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="cinder-scheduler" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050369 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" containerName="probe" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.050377 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" containerName="barbican-api" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.051318 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.054215 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.055544 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.076430 4708 scope.go:117] "RemoveContainer" containerID="a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.096568 4708 scope.go:117] "RemoveContainer" containerID="6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.113933 4708 scope.go:117] "RemoveContainer" containerID="a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.114274 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5\": container with ID starting with a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5 not found: ID does not exist" containerID="a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.114304 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5"} err="failed to get container status \"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5\": rpc error: code = NotFound desc = could not find container \"a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5\": container with ID starting with a83c139e6912ad4b0ba058e32d459e1480fe5c258799251c0b230bfd587106e5 not found: ID does not exist" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.114325 4708 scope.go:117] "RemoveContainer" containerID="6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9" Nov 25 05:55:47 crc kubenswrapper[4708]: E1125 05:55:47.114600 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9\": container with ID starting with 6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9 not found: ID does not exist" containerID="6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.114620 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9"} err="failed to get container status \"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9\": rpc error: code = NotFound desc = could not find container \"6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9\": container with ID starting with 6a57f0319b75364fafce9a13649de0ed4033cf6ffbf6bdee5de7ed1c981491c9 not found: ID does not exist" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144607 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-scripts\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144669 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1392ab-0788-442a-8664-1a674811e537-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144745 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144777 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144802 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.144848 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2jg5\" (UniqueName: \"kubernetes.io/projected/fc1392ab-0788-442a-8664-1a674811e537-kube-api-access-j2jg5\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246140 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-scripts\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246197 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1392ab-0788-442a-8664-1a674811e537-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246256 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246286 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246305 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.246347 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2jg5\" (UniqueName: \"kubernetes.io/projected/fc1392ab-0788-442a-8664-1a674811e537-kube-api-access-j2jg5\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.248150 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fc1392ab-0788-442a-8664-1a674811e537-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.252463 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.252471 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.252645 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-config-data\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.254131 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc1392ab-0788-442a-8664-1a674811e537-scripts\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.260830 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2jg5\" (UniqueName: \"kubernetes.io/projected/fc1392ab-0788-442a-8664-1a674811e537-kube-api-access-j2jg5\") pod \"cinder-scheduler-0\" (UID: \"fc1392ab-0788-442a-8664-1a674811e537\") " pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.370196 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.826543 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.985328 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fc1392ab-0788-442a-8664-1a674811e537","Type":"ContainerStarted","Data":"aa1bff927f6b18ef6d8277ba27e2c7ee7c69ace277c3039aa81c778c200408d3"} Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.990678 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerStarted","Data":"aa53e6fdfc3781cdf518becf08803b67da5d1c4664f917eb8a3492cca36a0b43"} Nov 25 05:55:47 crc kubenswrapper[4708]: I1125 05:55:47.992471 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fkw22" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="registry-server" containerID="cri-o://625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d" gracePeriod=2 Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.466446 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.488845 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content\") pod \"b03718e6-8cbd-4692-8ea1-89c9e6118469\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.488892 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities\") pod \"b03718e6-8cbd-4692-8ea1-89c9e6118469\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.489017 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vn8bg\" (UniqueName: \"kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg\") pod \"b03718e6-8cbd-4692-8ea1-89c9e6118469\" (UID: \"b03718e6-8cbd-4692-8ea1-89c9e6118469\") " Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.490635 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities" (OuterVolumeSpecName: "utilities") pod "b03718e6-8cbd-4692-8ea1-89c9e6118469" (UID: "b03718e6-8cbd-4692-8ea1-89c9e6118469"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.502718 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg" (OuterVolumeSpecName: "kube-api-access-vn8bg") pod "b03718e6-8cbd-4692-8ea1-89c9e6118469" (UID: "b03718e6-8cbd-4692-8ea1-89c9e6118469"). InnerVolumeSpecName "kube-api-access-vn8bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.530844 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b03718e6-8cbd-4692-8ea1-89c9e6118469" (UID: "b03718e6-8cbd-4692-8ea1-89c9e6118469"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.591294 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vn8bg\" (UniqueName: \"kubernetes.io/projected/b03718e6-8cbd-4692-8ea1-89c9e6118469-kube-api-access-vn8bg\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.591348 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.591368 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b03718e6-8cbd-4692-8ea1-89c9e6118469-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.905153 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6154c760-6930-4ceb-99d0-2aee2bf196ce" path="/var/lib/kubelet/pods/6154c760-6930-4ceb-99d0-2aee2bf196ce/volumes" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.906178 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3070821-6f6c-40fa-9755-e46c6651ee92" path="/var/lib/kubelet/pods/a3070821-6f6c-40fa-9755-e46c6651ee92/volumes" Nov 25 05:55:48 crc kubenswrapper[4708]: I1125 05:55:48.906822 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e208d102-59a4-4a0e-8041-84a4df5e4433" path="/var/lib/kubelet/pods/e208d102-59a4-4a0e-8041-84a4df5e4433/volumes" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.003324 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerStarted","Data":"ee7b0ef85ded26e5f4eb63f294b77c26cf7af9c18b8123c320b4f998e668f675"} Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.006015 4708 generic.go:334] "Generic (PLEG): container finished" podID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerID="625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d" exitCode=0 Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.006078 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fkw22" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.006087 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerDied","Data":"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d"} Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.006115 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fkw22" event={"ID":"b03718e6-8cbd-4692-8ea1-89c9e6118469","Type":"ContainerDied","Data":"5dc7cd5468e31e5bc20d088994c3d6bd3a038bd4c2eb72871f89b37c7317f817"} Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.006137 4708 scope.go:117] "RemoveContainer" containerID="625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.008166 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fc1392ab-0788-442a-8664-1a674811e537","Type":"ContainerStarted","Data":"153ee41b8902628862be394096dc74f70ff0800b60ad81b395aa8fdc059b2e9f"} Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.008194 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fc1392ab-0788-442a-8664-1a674811e537","Type":"ContainerStarted","Data":"32af015fdc50934fd4c3984f31d57190d9405f38217dd6bd9f2a95c8a8266351"} Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.026921 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.030973 4708 scope.go:117] "RemoveContainer" containerID="7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.040741 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fkw22"] Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.041131 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.041115796 podStartE2EDuration="2.041115796s" podCreationTimestamp="2025-11-25 05:55:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:55:49.031340594 +0000 UTC m=+890.440173980" watchObservedRunningTime="2025-11-25 05:55:49.041115796 +0000 UTC m=+890.449949182" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.054013 4708 scope.go:117] "RemoveContainer" containerID="0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.074734 4708 scope.go:117] "RemoveContainer" containerID="625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d" Nov 25 05:55:49 crc kubenswrapper[4708]: E1125 05:55:49.075376 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d\": container with ID starting with 625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d not found: ID does not exist" containerID="625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.075452 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d"} err="failed to get container status \"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d\": rpc error: code = NotFound desc = could not find container \"625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d\": container with ID starting with 625fdccb8588dcbf5b1d392f0c3a1308b25d251f41ab2db861f4c5b082904c7d not found: ID does not exist" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.075488 4708 scope.go:117] "RemoveContainer" containerID="7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103" Nov 25 05:55:49 crc kubenswrapper[4708]: E1125 05:55:49.076003 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103\": container with ID starting with 7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103 not found: ID does not exist" containerID="7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.076062 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103"} err="failed to get container status \"7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103\": rpc error: code = NotFound desc = could not find container \"7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103\": container with ID starting with 7cc86afd815b468565520c2b4015c7ed57e6eeb6b8488f4d8882b3e327f1d103 not found: ID does not exist" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.076104 4708 scope.go:117] "RemoveContainer" containerID="0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05" Nov 25 05:55:49 crc kubenswrapper[4708]: E1125 05:55:49.077069 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05\": container with ID starting with 0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05 not found: ID does not exist" containerID="0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05" Nov 25 05:55:49 crc kubenswrapper[4708]: I1125 05:55:49.077098 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05"} err="failed to get container status \"0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05\": rpc error: code = NotFound desc = could not find container \"0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05\": container with ID starting with 0c0d50e446c9ae78217ae4437935dd0ded59a7209ade8c89651380fcdc81ad05 not found: ID does not exist" Nov 25 05:55:50 crc kubenswrapper[4708]: I1125 05:55:50.906814 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" path="/var/lib/kubelet/pods/b03718e6-8cbd-4692-8ea1-89c9e6118469/volumes" Nov 25 05:55:51 crc kubenswrapper[4708]: I1125 05:55:51.039807 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerStarted","Data":"7024d63240f1c1f664e2b2fac2660498ca48dcdc3aea1bc4b7ff18b11caada3e"} Nov 25 05:55:51 crc kubenswrapper[4708]: I1125 05:55:51.040387 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:55:51 crc kubenswrapper[4708]: I1125 05:55:51.066276 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.978492496 podStartE2EDuration="7.066254651s" podCreationTimestamp="2025-11-25 05:55:44 +0000 UTC" firstStartedPulling="2025-11-25 05:55:45.844322801 +0000 UTC m=+887.253156186" lastFinishedPulling="2025-11-25 05:55:49.932084954 +0000 UTC m=+891.340918341" observedRunningTime="2025-11-25 05:55:51.058947669 +0000 UTC m=+892.467781065" watchObservedRunningTime="2025-11-25 05:55:51.066254651 +0000 UTC m=+892.475088036" Nov 25 05:55:52 crc kubenswrapper[4708]: I1125 05:55:52.371341 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 05:55:52 crc kubenswrapper[4708]: I1125 05:55:52.570633 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 05:55:56 crc kubenswrapper[4708]: I1125 05:55:56.385329 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:56 crc kubenswrapper[4708]: I1125 05:55:56.386284 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7b8ff6575b-8w7c4" Nov 25 05:55:57 crc kubenswrapper[4708]: I1125 05:55:57.621090 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 05:56:00 crc kubenswrapper[4708]: I1125 05:56:00.773646 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-db6d644f4-2hdvr" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.378683 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 05:56:04 crc kubenswrapper[4708]: E1125 05:56:04.380022 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="extract-utilities" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.380044 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="extract-utilities" Nov 25 05:56:04 crc kubenswrapper[4708]: E1125 05:56:04.380066 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="extract-content" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.380074 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="extract-content" Nov 25 05:56:04 crc kubenswrapper[4708]: E1125 05:56:04.380105 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="registry-server" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.380115 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="registry-server" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.384056 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03718e6-8cbd-4692-8ea1-89c9e6118469" containerName="registry-server" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.386006 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.393494 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.393776 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gzmbc" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.393950 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.415697 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.415914 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh8xc\" (UniqueName: \"kubernetes.io/projected/9c745e5e-026a-4659-8bfd-9c8939a234f4-kube-api-access-xh8xc\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.416106 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.416138 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.418800 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.518804 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.519013 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh8xc\" (UniqueName: \"kubernetes.io/projected/9c745e5e-026a-4659-8bfd-9c8939a234f4-kube-api-access-xh8xc\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.519204 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.519236 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.520144 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.525988 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.526022 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9c745e5e-026a-4659-8bfd-9c8939a234f4-openstack-config-secret\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.534667 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh8xc\" (UniqueName: \"kubernetes.io/projected/9c745e5e-026a-4659-8bfd-9c8939a234f4-kube-api-access-xh8xc\") pod \"openstackclient\" (UID: \"9c745e5e-026a-4659-8bfd-9c8939a234f4\") " pod="openstack/openstackclient" Nov 25 05:56:04 crc kubenswrapper[4708]: I1125 05:56:04.720440 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 05:56:05 crc kubenswrapper[4708]: I1125 05:56:05.140806 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 05:56:05 crc kubenswrapper[4708]: I1125 05:56:05.188827 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9c745e5e-026a-4659-8bfd-9c8939a234f4","Type":"ContainerStarted","Data":"80a422fbc0a2d2abf73931210cca913d94c342edccd6d339f04eceb37cacd63a"} Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.076830 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65b4dc74dc-ktwwc"] Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.079809 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.089007 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.089075 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.089198 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.096238 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65b4dc74dc-ktwwc"] Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.107917 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-public-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.107974 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-combined-ca-bundle\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.108008 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-run-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.108056 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-internal-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.108093 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-log-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.108181 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-etc-swift\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.109076 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-config-data\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.109145 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jstcp\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-kube-api-access-jstcp\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210467 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-etc-swift\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210535 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-config-data\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210580 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jstcp\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-kube-api-access-jstcp\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210616 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-public-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210648 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-combined-ca-bundle\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210672 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-run-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.210716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-internal-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.211700 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-run-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.212021 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-log-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.212314 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-log-httpd\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.217995 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-config-data\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.219260 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-etc-swift\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.219998 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-internal-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.220159 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-public-tls-certs\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.220713 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-combined-ca-bundle\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.224459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jstcp\" (UniqueName: \"kubernetes.io/projected/34548024-cb82-4fa9-ace7-9e41b3ab2c1d-kube-api-access-jstcp\") pod \"swift-proxy-65b4dc74dc-ktwwc\" (UID: \"34548024-cb82-4fa9-ace7-9e41b3ab2c1d\") " pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.422566 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:09 crc kubenswrapper[4708]: I1125 05:56:09.938998 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65b4dc74dc-ktwwc"] Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.249074 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" event={"ID":"34548024-cb82-4fa9-ace7-9e41b3ab2c1d","Type":"ContainerStarted","Data":"86788fb9afe00ed2620366bcfdab469797d4055233b642f182e4681d8625e2e1"} Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.249287 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" event={"ID":"34548024-cb82-4fa9-ace7-9e41b3ab2c1d","Type":"ContainerStarted","Data":"f92ba1ec2c5a42184b0845ed27261893d6816701d488a8bb5d4cf429de2ad0a6"} Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.518786 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.519098 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-central-agent" containerID="cri-o://11e7c10b8784262fe65ec4e3c68b654129ff238dffe7b4525c2fc1ca2d0277e3" gracePeriod=30 Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.519434 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="proxy-httpd" containerID="cri-o://7024d63240f1c1f664e2b2fac2660498ca48dcdc3aea1bc4b7ff18b11caada3e" gracePeriod=30 Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.519464 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-notification-agent" containerID="cri-o://aa53e6fdfc3781cdf518becf08803b67da5d1c4664f917eb8a3492cca36a0b43" gracePeriod=30 Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.519571 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="sg-core" containerID="cri-o://ee7b0ef85ded26e5f4eb63f294b77c26cf7af9c18b8123c320b4f998e668f675" gracePeriod=30 Nov 25 05:56:10 crc kubenswrapper[4708]: I1125 05:56:10.529247 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.266583 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerID="7024d63240f1c1f664e2b2fac2660498ca48dcdc3aea1bc4b7ff18b11caada3e" exitCode=0 Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.266950 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerID="ee7b0ef85ded26e5f4eb63f294b77c26cf7af9c18b8123c320b4f998e668f675" exitCode=2 Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.266963 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerID="11e7c10b8784262fe65ec4e3c68b654129ff238dffe7b4525c2fc1ca2d0277e3" exitCode=0 Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.266678 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerDied","Data":"7024d63240f1c1f664e2b2fac2660498ca48dcdc3aea1bc4b7ff18b11caada3e"} Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.267087 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerDied","Data":"ee7b0ef85ded26e5f4eb63f294b77c26cf7af9c18b8123c320b4f998e668f675"} Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.267108 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerDied","Data":"11e7c10b8784262fe65ec4e3c68b654129ff238dffe7b4525c2fc1ca2d0277e3"} Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.272313 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" event={"ID":"34548024-cb82-4fa9-ace7-9e41b3ab2c1d","Type":"ContainerStarted","Data":"cdce19a8af20db19dfe5b8b4e382b4b7f92dbe14a64d53146889239a3ad71765"} Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.272571 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:11 crc kubenswrapper[4708]: I1125 05:56:11.291783 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" podStartSLOduration=2.2917635020000002 podStartE2EDuration="2.291763502s" podCreationTimestamp="2025-11-25 05:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:11.28917818 +0000 UTC m=+912.698011566" watchObservedRunningTime="2025-11-25 05:56:11.291763502 +0000 UTC m=+912.700596889" Nov 25 05:56:12 crc kubenswrapper[4708]: I1125 05:56:12.298758 4708 generic.go:334] "Generic (PLEG): container finished" podID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerID="aa53e6fdfc3781cdf518becf08803b67da5d1c4664f917eb8a3492cca36a0b43" exitCode=0 Nov 25 05:56:12 crc kubenswrapper[4708]: I1125 05:56:12.298843 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerDied","Data":"aa53e6fdfc3781cdf518becf08803b67da5d1c4664f917eb8a3492cca36a0b43"} Nov 25 05:56:12 crc kubenswrapper[4708]: I1125 05:56:12.298997 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:14 crc kubenswrapper[4708]: I1125 05:56:14.357539 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:56:14 crc kubenswrapper[4708]: I1125 05:56:14.357633 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.233202 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-5bf76"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.234662 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.247642 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5bf76"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.339822 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": dial tcp 10.217.0.160:3000: connect: connection refused" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.344853 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0add-account-create-cxk7s"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.346262 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.347123 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.347216 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r9q4\" (UniqueName: \"kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.361882 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.382879 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0add-account-create-cxk7s"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.441810 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jms52"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.443202 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.452502 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.452561 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.452614 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r9q4\" (UniqueName: \"kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.452666 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k2pg\" (UniqueName: \"kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.453292 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.479287 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r9q4\" (UniqueName: \"kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4\") pod \"nova-api-db-create-5bf76\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.493608 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jms52"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.532773 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-k52w2"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.534228 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.549954 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-53a6-account-create-fwcmk"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.551228 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.552713 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.554707 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.554834 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k2pg\" (UniqueName: \"kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.554937 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkxt2\" (UniqueName: \"kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.554996 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.556239 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.559693 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.561675 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-k52w2"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.573458 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-53a6-account-create-fwcmk"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.581341 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k2pg\" (UniqueName: \"kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg\") pod \"nova-api-0add-account-create-cxk7s\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657384 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l75jr\" (UniqueName: \"kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657478 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657538 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nzfq\" (UniqueName: \"kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657574 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657810 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.657931 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkxt2\" (UniqueName: \"kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.658636 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.671968 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkxt2\" (UniqueName: \"kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2\") pod \"nova-cell0-db-create-jms52\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.681571 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.718316 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-084d-account-create-gx5q9"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.719793 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.721948 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.734797 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-084d-account-create-gx5q9"] Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.760677 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l75jr\" (UniqueName: \"kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.760801 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nzfq\" (UniqueName: \"kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.760841 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.760897 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.761631 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.761809 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.771540 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.775109 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l75jr\" (UniqueName: \"kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr\") pod \"nova-cell1-db-create-k52w2\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.775364 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nzfq\" (UniqueName: \"kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq\") pod \"nova-cell0-53a6-account-create-fwcmk\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.862171 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.862215 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rz42\" (UniqueName: \"kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.909093 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.912150 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.965132 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.965279 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rz42\" (UniqueName: \"kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.967446 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:15 crc kubenswrapper[4708]: I1125 05:56:15.982316 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rz42\" (UniqueName: \"kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42\") pod \"nova-cell1-084d-account-create-gx5q9\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:16 crc kubenswrapper[4708]: I1125 05:56:16.037064 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.316318 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.383297 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3ce59277-ab22-49bd-b22a-c3769e2e3333","Type":"ContainerDied","Data":"4ca2eba6fe9e1a67e738f51be3bd8a0b24eb7202f80f1866914844b5b6bf1158"} Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.383349 4708 scope.go:117] "RemoveContainer" containerID="7024d63240f1c1f664e2b2fac2660498ca48dcdc3aea1bc4b7ff18b11caada3e" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.383488 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.386987 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9c745e5e-026a-4659-8bfd-9c8939a234f4","Type":"ContainerStarted","Data":"fb90193c4e73a4ef53fdb601d3771c3cd9d5e0f8a182735b4031d5b93839c7d2"} Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407091 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407129 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407179 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407216 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407362 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407416 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7rrr\" (UniqueName: \"kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.407494 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd\") pod \"3ce59277-ab22-49bd-b22a-c3769e2e3333\" (UID: \"3ce59277-ab22-49bd-b22a-c3769e2e3333\") " Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.410028 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.411587 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.414735 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr" (OuterVolumeSpecName: "kube-api-access-g7rrr") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "kube-api-access-g7rrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.415007 4708 scope.go:117] "RemoveContainer" containerID="ee7b0ef85ded26e5f4eb63f294b77c26cf7af9c18b8123c320b4f998e668f675" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.416253 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.418435335 podStartE2EDuration="13.416235177s" podCreationTimestamp="2025-11-25 05:56:04 +0000 UTC" firstStartedPulling="2025-11-25 05:56:05.141643569 +0000 UTC m=+906.550476955" lastFinishedPulling="2025-11-25 05:56:17.139443412 +0000 UTC m=+918.548276797" observedRunningTime="2025-11-25 05:56:17.405373617 +0000 UTC m=+918.814207003" watchObservedRunningTime="2025-11-25 05:56:17.416235177 +0000 UTC m=+918.825068563" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.421550 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts" (OuterVolumeSpecName: "scripts") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.443697 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.452674 4708 scope.go:117] "RemoveContainer" containerID="aa53e6fdfc3781cdf518becf08803b67da5d1c4664f917eb8a3492cca36a0b43" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.474384 4708 scope.go:117] "RemoveContainer" containerID="11e7c10b8784262fe65ec4e3c68b654129ff238dffe7b4525c2fc1ca2d0277e3" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.510450 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.510478 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7rrr\" (UniqueName: \"kubernetes.io/projected/3ce59277-ab22-49bd-b22a-c3769e2e3333-kube-api-access-g7rrr\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.510492 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.510501 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.510510 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3ce59277-ab22-49bd-b22a-c3769e2e3333-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.540568 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data" (OuterVolumeSpecName: "config-data") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.541345 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ce59277-ab22-49bd-b22a-c3769e2e3333" (UID: "3ce59277-ab22-49bd-b22a-c3769e2e3333"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.612778 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.612838 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ce59277-ab22-49bd-b22a-c3769e2e3333-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.632206 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-084d-account-create-gx5q9"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.698506 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0add-account-create-cxk7s"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.770830 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.777496 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.786350 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-k52w2"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.797604 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:17 crc kubenswrapper[4708]: E1125 05:56:17.798040 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-notification-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798060 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-notification-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: E1125 05:56:17.798073 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="sg-core" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798080 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="sg-core" Nov 25 05:56:17 crc kubenswrapper[4708]: E1125 05:56:17.798110 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-central-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798116 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-central-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: E1125 05:56:17.798126 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="proxy-httpd" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798132 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="proxy-httpd" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798324 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="sg-core" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798342 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-notification-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798352 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="ceilometer-central-agent" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.798368 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" containerName="proxy-httpd" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.807265 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.810069 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.810247 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.814937 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-5bf76"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816269 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lm22\" (UniqueName: \"kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816317 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816352 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816385 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816428 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816484 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.816514 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.830556 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jms52"] Nov 25 05:56:17 crc kubenswrapper[4708]: W1125 05:56:17.849915 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12614737_33a5_4a01_bd4a_91e9e41e0022.slice/crio-1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59 WatchSource:0}: Error finding container 1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59: Status 404 returned error can't find the container with id 1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59 Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.850267 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-53a6-account-create-fwcmk"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.857632 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921479 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921586 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921643 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921814 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.921884 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.922067 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.922180 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lm22\" (UniqueName: \"kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.923127 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.927173 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.928445 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.929018 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.930854 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:17 crc kubenswrapper[4708]: I1125 05:56:17.938859 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lm22\" (UniqueName: \"kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22\") pod \"ceilometer-0\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " pod="openstack/ceilometer-0" Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.147003 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.395684 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.396229 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-log" containerID="cri-o://943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c" gracePeriod=30 Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.396452 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-httpd" containerID="cri-o://ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0" gracePeriod=30 Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.410857 4708 generic.go:334] "Generic (PLEG): container finished" podID="099277b8-d95e-41b6-b0d4-33f6346752c5" containerID="e29f0bc44d62fee9a60e174d60c2dac676316f4cd8170eb024f4df3196ad31e8" exitCode=0 Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.411283 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0add-account-create-cxk7s" event={"ID":"099277b8-d95e-41b6-b0d4-33f6346752c5","Type":"ContainerDied","Data":"e29f0bc44d62fee9a60e174d60c2dac676316f4cd8170eb024f4df3196ad31e8"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.411313 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0add-account-create-cxk7s" event={"ID":"099277b8-d95e-41b6-b0d4-33f6346752c5","Type":"ContainerStarted","Data":"d1b3e79aca9ac7db7189125b0be1edfce8da247098b37049d076c78879c08e57"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.412921 4708 generic.go:334] "Generic (PLEG): container finished" podID="12614737-33a5-4a01-bd4a-91e9e41e0022" containerID="946fcc0147e7f344e51461ce86a2b8012291b42ea3d6b1254bfa7b921e349f55" exitCode=0 Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.412984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bf76" event={"ID":"12614737-33a5-4a01-bd4a-91e9e41e0022","Type":"ContainerDied","Data":"946fcc0147e7f344e51461ce86a2b8012291b42ea3d6b1254bfa7b921e349f55"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.413013 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bf76" event={"ID":"12614737-33a5-4a01-bd4a-91e9e41e0022","Type":"ContainerStarted","Data":"1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.414646 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jms52" event={"ID":"99663ec8-6afb-455e-8fbb-ef9c015b6c60","Type":"ContainerStarted","Data":"36369ecde8c56951a3357e3f969cdfacab7e267a99860b9d9ef8e816a9b59362"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.414670 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jms52" event={"ID":"99663ec8-6afb-455e-8fbb-ef9c015b6c60","Type":"ContainerStarted","Data":"e53d9e627d1de7b1ec0a7513fd5fea42b796ede1ee642ad0092d5f1fc00731a4"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.420449 4708 generic.go:334] "Generic (PLEG): container finished" podID="a9957cda-e4de-40f4-87fc-ce0dfd166505" containerID="237199b481c43a69bbff1a05921d2624cc211bf25c7f1138db9af1168796e82a" exitCode=0 Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.420493 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084d-account-create-gx5q9" event={"ID":"a9957cda-e4de-40f4-87fc-ce0dfd166505","Type":"ContainerDied","Data":"237199b481c43a69bbff1a05921d2624cc211bf25c7f1138db9af1168796e82a"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.420511 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084d-account-create-gx5q9" event={"ID":"a9957cda-e4de-40f4-87fc-ce0dfd166505","Type":"ContainerStarted","Data":"ea5f7a0fbdcd921cfd124978ada898dffc1f7fd5a23dd07a7003322f3313e6f0"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.421580 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53a6-account-create-fwcmk" event={"ID":"4b06819a-2b30-4c4a-aa31-97e1482aeab1","Type":"ContainerStarted","Data":"9b99ce75bdfdeb5995cbb6a828c83d5dadb2dfad727065db138261823ea11f50"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.421601 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53a6-account-create-fwcmk" event={"ID":"4b06819a-2b30-4c4a-aa31-97e1482aeab1","Type":"ContainerStarted","Data":"d1bb20af85f0f0a9218b6adbc30d06d33d8432c53190e1f0aa889b5ecef4517a"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.423101 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-k52w2" event={"ID":"a91604ea-9b5a-4e39-96ec-abb8302360a0","Type":"ContainerStarted","Data":"4969ab9724ca3c3f14130a3efdd9fa73c081d7e796b5efabf94f5301c2c5259e"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.423133 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-k52w2" event={"ID":"a91604ea-9b5a-4e39-96ec-abb8302360a0","Type":"ContainerStarted","Data":"90d92186f26e2b219bef88421d8d77ba44574209a7d53288b1af444d1c32effc"} Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.495419 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-k52w2" podStartSLOduration=3.495390601 podStartE2EDuration="3.495390601s" podCreationTimestamp="2025-11-25 05:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:18.481171172 +0000 UTC m=+919.890004559" watchObservedRunningTime="2025-11-25 05:56:18.495390601 +0000 UTC m=+919.904223988" Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.520795 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-jms52" podStartSLOduration=3.520771886 podStartE2EDuration="3.520771886s" podCreationTimestamp="2025-11-25 05:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:18.513032992 +0000 UTC m=+919.921866377" watchObservedRunningTime="2025-11-25 05:56:18.520771886 +0000 UTC m=+919.929605272" Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.533072 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-53a6-account-create-fwcmk" podStartSLOduration=3.53305187 podStartE2EDuration="3.53305187s" podCreationTimestamp="2025-11-25 05:56:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:18.528631359 +0000 UTC m=+919.937464745" watchObservedRunningTime="2025-11-25 05:56:18.53305187 +0000 UTC m=+919.941885256" Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.758882 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:18 crc kubenswrapper[4708]: I1125 05:56:18.905387 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ce59277-ab22-49bd-b22a-c3769e2e3333" path="/var/lib/kubelet/pods/3ce59277-ab22-49bd-b22a-c3769e2e3333/volumes" Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.324219 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.324490 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-log" containerID="cri-o://0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7" gracePeriod=30 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.324592 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-httpd" containerID="cri-o://37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e" gracePeriod=30 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.437902 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.445649 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65b4dc74dc-ktwwc" Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.445781 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerStarted","Data":"815602f04cf2347e530f6e595c8d7be7178ca77b9b2e46db1d5f0007ea2cc349"} Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.447719 4708 generic.go:334] "Generic (PLEG): container finished" podID="a91604ea-9b5a-4e39-96ec-abb8302360a0" containerID="4969ab9724ca3c3f14130a3efdd9fa73c081d7e796b5efabf94f5301c2c5259e" exitCode=0 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.447782 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-k52w2" event={"ID":"a91604ea-9b5a-4e39-96ec-abb8302360a0","Type":"ContainerDied","Data":"4969ab9724ca3c3f14130a3efdd9fa73c081d7e796b5efabf94f5301c2c5259e"} Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.453794 4708 generic.go:334] "Generic (PLEG): container finished" podID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerID="943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c" exitCode=143 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.453846 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerDied","Data":"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c"} Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.456849 4708 generic.go:334] "Generic (PLEG): container finished" podID="99663ec8-6afb-455e-8fbb-ef9c015b6c60" containerID="36369ecde8c56951a3357e3f969cdfacab7e267a99860b9d9ef8e816a9b59362" exitCode=0 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.457039 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jms52" event={"ID":"99663ec8-6afb-455e-8fbb-ef9c015b6c60","Type":"ContainerDied","Data":"36369ecde8c56951a3357e3f969cdfacab7e267a99860b9d9ef8e816a9b59362"} Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.467839 4708 generic.go:334] "Generic (PLEG): container finished" podID="4b06819a-2b30-4c4a-aa31-97e1482aeab1" containerID="9b99ce75bdfdeb5995cbb6a828c83d5dadb2dfad727065db138261823ea11f50" exitCode=0 Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.468125 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53a6-account-create-fwcmk" event={"ID":"4b06819a-2b30-4c4a-aa31-97e1482aeab1","Type":"ContainerDied","Data":"9b99ce75bdfdeb5995cbb6a828c83d5dadb2dfad727065db138261823ea11f50"} Nov 25 05:56:19 crc kubenswrapper[4708]: I1125 05:56:19.908677 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.025081 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.029682 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.076336 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts\") pod \"099277b8-d95e-41b6-b0d4-33f6346752c5\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.076829 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7k2pg\" (UniqueName: \"kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg\") pod \"099277b8-d95e-41b6-b0d4-33f6346752c5\" (UID: \"099277b8-d95e-41b6-b0d4-33f6346752c5\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.077794 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "099277b8-d95e-41b6-b0d4-33f6346752c5" (UID: "099277b8-d95e-41b6-b0d4-33f6346752c5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.078125 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/099277b8-d95e-41b6-b0d4-33f6346752c5-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.090130 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg" (OuterVolumeSpecName: "kube-api-access-7k2pg") pod "099277b8-d95e-41b6-b0d4-33f6346752c5" (UID: "099277b8-d95e-41b6-b0d4-33f6346752c5"). InnerVolumeSpecName "kube-api-access-7k2pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.179645 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5r9q4\" (UniqueName: \"kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4\") pod \"12614737-33a5-4a01-bd4a-91e9e41e0022\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.179907 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts\") pod \"a9957cda-e4de-40f4-87fc-ce0dfd166505\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.180004 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts\") pod \"12614737-33a5-4a01-bd4a-91e9e41e0022\" (UID: \"12614737-33a5-4a01-bd4a-91e9e41e0022\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.180074 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rz42\" (UniqueName: \"kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42\") pod \"a9957cda-e4de-40f4-87fc-ce0dfd166505\" (UID: \"a9957cda-e4de-40f4-87fc-ce0dfd166505\") " Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.180411 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a9957cda-e4de-40f4-87fc-ce0dfd166505" (UID: "a9957cda-e4de-40f4-87fc-ce0dfd166505"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.180551 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "12614737-33a5-4a01-bd4a-91e9e41e0022" (UID: "12614737-33a5-4a01-bd4a-91e9e41e0022"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.181308 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7k2pg\" (UniqueName: \"kubernetes.io/projected/099277b8-d95e-41b6-b0d4-33f6346752c5-kube-api-access-7k2pg\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.181340 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a9957cda-e4de-40f4-87fc-ce0dfd166505-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.181351 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/12614737-33a5-4a01-bd4a-91e9e41e0022-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.183889 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4" (OuterVolumeSpecName: "kube-api-access-5r9q4") pod "12614737-33a5-4a01-bd4a-91e9e41e0022" (UID: "12614737-33a5-4a01-bd4a-91e9e41e0022"). InnerVolumeSpecName "kube-api-access-5r9q4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.186684 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42" (OuterVolumeSpecName: "kube-api-access-5rz42") pod "a9957cda-e4de-40f4-87fc-ce0dfd166505" (UID: "a9957cda-e4de-40f4-87fc-ce0dfd166505"). InnerVolumeSpecName "kube-api-access-5rz42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.283968 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rz42\" (UniqueName: \"kubernetes.io/projected/a9957cda-e4de-40f4-87fc-ce0dfd166505-kube-api-access-5rz42\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.284002 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5r9q4\" (UniqueName: \"kubernetes.io/projected/12614737-33a5-4a01-bd4a-91e9e41e0022-kube-api-access-5r9q4\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.483952 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0add-account-create-cxk7s" event={"ID":"099277b8-d95e-41b6-b0d4-33f6346752c5","Type":"ContainerDied","Data":"d1b3e79aca9ac7db7189125b0be1edfce8da247098b37049d076c78879c08e57"} Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.484210 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1b3e79aca9ac7db7189125b0be1edfce8da247098b37049d076c78879c08e57" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.484278 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0add-account-create-cxk7s" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.489897 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-5bf76" event={"ID":"12614737-33a5-4a01-bd4a-91e9e41e0022","Type":"ContainerDied","Data":"1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59"} Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.489932 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b7533433c1c350d8073beb12c77cb14443c2a4a421a8c9828011ae7798f7f59" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.489988 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-5bf76" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.497768 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-084d-account-create-gx5q9" event={"ID":"a9957cda-e4de-40f4-87fc-ce0dfd166505","Type":"ContainerDied","Data":"ea5f7a0fbdcd921cfd124978ada898dffc1f7fd5a23dd07a7003322f3313e6f0"} Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.497802 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea5f7a0fbdcd921cfd124978ada898dffc1f7fd5a23dd07a7003322f3313e6f0" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.497860 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-084d-account-create-gx5q9" Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.500760 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerStarted","Data":"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369"} Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.500809 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerStarted","Data":"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846"} Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.502961 4708 generic.go:334] "Generic (PLEG): container finished" podID="493b9467-e848-4aae-a5a5-5d699811e40e" containerID="0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7" exitCode=143 Nov 25 05:56:20 crc kubenswrapper[4708]: I1125 05:56:20.503047 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerDied","Data":"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7"} Nov 25 05:56:20 crc kubenswrapper[4708]: E1125 05:56:20.923881 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099277b8_d95e_41b6_b0d4_33f6346752c5.slice/crio-d1b3e79aca9ac7db7189125b0be1edfce8da247098b37049d076c78879c08e57\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9957cda_e4de_40f4_87fc_ce0dfd166505.slice/crio-ea5f7a0fbdcd921cfd124978ada898dffc1f7fd5a23dd07a7003322f3313e6f0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod099277b8_d95e_41b6_b0d4_33f6346752c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12614737_33a5_4a01_bd4a_91e9e41e0022.slice\": RecentStats: unable to find data in memory cache]" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.086151 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.111382 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.114265 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.239412 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l75jr\" (UniqueName: \"kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr\") pod \"a91604ea-9b5a-4e39-96ec-abb8302360a0\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.239818 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nzfq\" (UniqueName: \"kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq\") pod \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.239878 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts\") pod \"a91604ea-9b5a-4e39-96ec-abb8302360a0\" (UID: \"a91604ea-9b5a-4e39-96ec-abb8302360a0\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.240279 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts\") pod \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\" (UID: \"4b06819a-2b30-4c4a-aa31-97e1482aeab1\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.240338 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkxt2\" (UniqueName: \"kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2\") pod \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.240405 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts\") pod \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\" (UID: \"99663ec8-6afb-455e-8fbb-ef9c015b6c60\") " Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.240467 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a91604ea-9b5a-4e39-96ec-abb8302360a0" (UID: "a91604ea-9b5a-4e39-96ec-abb8302360a0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.240745 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4b06819a-2b30-4c4a-aa31-97e1482aeab1" (UID: "4b06819a-2b30-4c4a-aa31-97e1482aeab1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.241270 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99663ec8-6afb-455e-8fbb-ef9c015b6c60" (UID: "99663ec8-6afb-455e-8fbb-ef9c015b6c60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.241383 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4b06819a-2b30-4c4a-aa31-97e1482aeab1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.241409 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99663ec8-6afb-455e-8fbb-ef9c015b6c60-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.241420 4708 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a91604ea-9b5a-4e39-96ec-abb8302360a0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.246996 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2" (OuterVolumeSpecName: "kube-api-access-nkxt2") pod "99663ec8-6afb-455e-8fbb-ef9c015b6c60" (UID: "99663ec8-6afb-455e-8fbb-ef9c015b6c60"). InnerVolumeSpecName "kube-api-access-nkxt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.247086 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq" (OuterVolumeSpecName: "kube-api-access-9nzfq") pod "4b06819a-2b30-4c4a-aa31-97e1482aeab1" (UID: "4b06819a-2b30-4c4a-aa31-97e1482aeab1"). InnerVolumeSpecName "kube-api-access-9nzfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.247112 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr" (OuterVolumeSpecName: "kube-api-access-l75jr") pod "a91604ea-9b5a-4e39-96ec-abb8302360a0" (UID: "a91604ea-9b5a-4e39-96ec-abb8302360a0"). InnerVolumeSpecName "kube-api-access-l75jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.343331 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkxt2\" (UniqueName: \"kubernetes.io/projected/99663ec8-6afb-455e-8fbb-ef9c015b6c60-kube-api-access-nkxt2\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.343363 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l75jr\" (UniqueName: \"kubernetes.io/projected/a91604ea-9b5a-4e39-96ec-abb8302360a0-kube-api-access-l75jr\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.343375 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nzfq\" (UniqueName: \"kubernetes.io/projected/4b06819a-2b30-4c4a-aa31-97e1482aeab1-kube-api-access-9nzfq\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.513491 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jms52" event={"ID":"99663ec8-6afb-455e-8fbb-ef9c015b6c60","Type":"ContainerDied","Data":"e53d9e627d1de7b1ec0a7513fd5fea42b796ede1ee642ad0092d5f1fc00731a4"} Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.513549 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e53d9e627d1de7b1ec0a7513fd5fea42b796ede1ee642ad0092d5f1fc00731a4" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.513545 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jms52" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.516416 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-53a6-account-create-fwcmk" event={"ID":"4b06819a-2b30-4c4a-aa31-97e1482aeab1","Type":"ContainerDied","Data":"d1bb20af85f0f0a9218b6adbc30d06d33d8432c53190e1f0aa889b5ecef4517a"} Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.516459 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1bb20af85f0f0a9218b6adbc30d06d33d8432c53190e1f0aa889b5ecef4517a" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.516501 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-53a6-account-create-fwcmk" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.520812 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerStarted","Data":"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6"} Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.522311 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-k52w2" event={"ID":"a91604ea-9b5a-4e39-96ec-abb8302360a0","Type":"ContainerDied","Data":"90d92186f26e2b219bef88421d8d77ba44574209a7d53288b1af444d1c32effc"} Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.522357 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="90d92186f26e2b219bef88421d8d77ba44574209a7d53288b1af444d1c32effc" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.522395 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-k52w2" Nov 25 05:56:21 crc kubenswrapper[4708]: I1125 05:56:21.984780 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057089 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057175 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057206 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057245 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75p4k\" (UniqueName: \"kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057405 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.057786 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs" (OuterVolumeSpecName: "logs") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.058192 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.058220 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.058451 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs\") pod \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\" (UID: \"f69ca19b-b4ff-4fce-b188-8dbda9d3a301\") " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.058674 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.059207 4708 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.059226 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.064737 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts" (OuterVolumeSpecName: "scripts") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.065674 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.073603 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k" (OuterVolumeSpecName: "kube-api-access-75p4k") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "kube-api-access-75p4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.112451 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.112616 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.123339 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data" (OuterVolumeSpecName: "config-data") pod "f69ca19b-b4ff-4fce-b188-8dbda9d3a301" (UID: "f69ca19b-b4ff-4fce-b188-8dbda9d3a301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161205 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161236 4708 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161249 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161260 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161269 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75p4k\" (UniqueName: \"kubernetes.io/projected/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-kube-api-access-75p4k\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.161278 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f69ca19b-b4ff-4fce-b188-8dbda9d3a301-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.178724 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.263112 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.547074 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerStarted","Data":"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987"} Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.551773 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.554176 4708 generic.go:334] "Generic (PLEG): container finished" podID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerID="ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0" exitCode=0 Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.554233 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.554243 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerDied","Data":"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0"} Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.555596 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f69ca19b-b4ff-4fce-b188-8dbda9d3a301","Type":"ContainerDied","Data":"3401e4e693bc1228f1814338357cdfa36aeadc9a9e85c74f9460e3f2e6387a3d"} Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.555619 4708 scope.go:117] "RemoveContainer" containerID="ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.583310 4708 scope.go:117] "RemoveContainer" containerID="943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.583542 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.05428363 podStartE2EDuration="5.583511021s" podCreationTimestamp="2025-11-25 05:56:17 +0000 UTC" firstStartedPulling="2025-11-25 05:56:18.752684732 +0000 UTC m=+920.161518117" lastFinishedPulling="2025-11-25 05:56:22.281912122 +0000 UTC m=+923.690745508" observedRunningTime="2025-11-25 05:56:22.577291248 +0000 UTC m=+923.986124635" watchObservedRunningTime="2025-11-25 05:56:22.583511021 +0000 UTC m=+923.992344407" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.601134 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.615515 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.617706 4708 scope.go:117] "RemoveContainer" containerID="ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.623640 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0\": container with ID starting with ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0 not found: ID does not exist" containerID="ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.623772 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0"} err="failed to get container status \"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0\": rpc error: code = NotFound desc = could not find container \"ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0\": container with ID starting with ce5bcf3376220bb4ad5045f7b62c5b9c251c6f82ec217083ed60f5c61348fec0 not found: ID does not exist" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.623866 4708 scope.go:117] "RemoveContainer" containerID="943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.625397 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c\": container with ID starting with 943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c not found: ID does not exist" containerID="943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.625509 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c"} err="failed to get container status \"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c\": rpc error: code = NotFound desc = could not find container \"943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c\": container with ID starting with 943cf4ccf2f3308e2c5ee18bf559d251ddeb2b442bc97e62951c9b4d9070625c not found: ID does not exist" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.634659 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635151 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9957cda-e4de-40f4-87fc-ce0dfd166505" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635172 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9957cda-e4de-40f4-87fc-ce0dfd166505" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635185 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b06819a-2b30-4c4a-aa31-97e1482aeab1" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635191 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b06819a-2b30-4c4a-aa31-97e1482aeab1" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635201 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="099277b8-d95e-41b6-b0d4-33f6346752c5" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635207 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="099277b8-d95e-41b6-b0d4-33f6346752c5" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635213 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-httpd" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635219 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-httpd" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635229 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12614737-33a5-4a01-bd4a-91e9e41e0022" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635234 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="12614737-33a5-4a01-bd4a-91e9e41e0022" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635242 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99663ec8-6afb-455e-8fbb-ef9c015b6c60" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635248 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="99663ec8-6afb-455e-8fbb-ef9c015b6c60" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635280 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-log" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635287 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-log" Nov 25 05:56:22 crc kubenswrapper[4708]: E1125 05:56:22.635299 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a91604ea-9b5a-4e39-96ec-abb8302360a0" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635305 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a91604ea-9b5a-4e39-96ec-abb8302360a0" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635462 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a91604ea-9b5a-4e39-96ec-abb8302360a0" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635474 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9957cda-e4de-40f4-87fc-ce0dfd166505" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635482 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="12614737-33a5-4a01-bd4a-91e9e41e0022" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635490 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="099277b8-d95e-41b6-b0d4-33f6346752c5" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635497 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="99663ec8-6afb-455e-8fbb-ef9c015b6c60" containerName="mariadb-database-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635511 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-log" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635540 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" containerName="glance-httpd" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.635549 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b06819a-2b30-4c4a-aa31-97e1482aeab1" containerName="mariadb-account-create" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.636540 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.643024 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.643155 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.653089 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.772780 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.772819 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773045 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773095 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8xp9\" (UniqueName: \"kubernetes.io/projected/0af2b359-a58e-4bc7-b497-40368df40ee1-kube-api-access-z8xp9\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773162 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-logs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773198 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773216 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.773350 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875484 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875842 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875868 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875921 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875946 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8xp9\" (UniqueName: \"kubernetes.io/projected/0af2b359-a58e-4bc7-b497-40368df40ee1-kube-api-access-z8xp9\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875968 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-logs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.875990 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.876005 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.876983 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.877077 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0af2b359-a58e-4bc7-b497-40368df40ee1-logs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.877260 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.884321 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-scripts\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.884734 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-config-data\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.885751 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.886610 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0af2b359-a58e-4bc7-b497-40368df40ee1-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.894970 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8xp9\" (UniqueName: \"kubernetes.io/projected/0af2b359-a58e-4bc7-b497-40368df40ee1-kube-api-access-z8xp9\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.902939 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"0af2b359-a58e-4bc7-b497-40368df40ee1\") " pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.916379 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f69ca19b-b4ff-4fce-b188-8dbda9d3a301" path="/var/lib/kubelet/pods/f69ca19b-b4ff-4fce-b188-8dbda9d3a301/volumes" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.957231 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Nov 25 05:56:22 crc kubenswrapper[4708]: I1125 05:56:22.980260 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.033659 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081102 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59g92\" (UniqueName: \"kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081147 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081182 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081224 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081242 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081258 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081271 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.081287 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle\") pod \"493b9467-e848-4aae-a5a5-5d699811e40e\" (UID: \"493b9467-e848-4aae-a5a5-5d699811e40e\") " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.084673 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.084981 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs" (OuterVolumeSpecName: "logs") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.091369 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92" (OuterVolumeSpecName: "kube-api-access-59g92") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "kube-api-access-59g92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.093496 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts" (OuterVolumeSpecName: "scripts") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.094373 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.132814 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.136511 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data" (OuterVolumeSpecName: "config-data") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.183600 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "493b9467-e848-4aae-a5a5-5d699811e40e" (UID: "493b9467-e848-4aae-a5a5-5d699811e40e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184637 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59g92\" (UniqueName: \"kubernetes.io/projected/493b9467-e848-4aae-a5a5-5d699811e40e-kube-api-access-59g92\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184672 4708 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184686 4708 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184723 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184733 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184742 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184754 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/493b9467-e848-4aae-a5a5-5d699811e40e-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.184763 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/493b9467-e848-4aae-a5a5-5d699811e40e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.253277 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.287030 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.536323 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.575137 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0af2b359-a58e-4bc7-b497-40368df40ee1","Type":"ContainerStarted","Data":"234c7628dbb5d638ae2398372c432bc2ffaff2f814cf5cee50eda177b21bf2c3"} Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.582153 4708 generic.go:334] "Generic (PLEG): container finished" podID="493b9467-e848-4aae-a5a5-5d699811e40e" containerID="37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e" exitCode=0 Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.582308 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerDied","Data":"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e"} Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.582392 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"493b9467-e848-4aae-a5a5-5d699811e40e","Type":"ContainerDied","Data":"ac1e9e7872f1b74c8133f0acfcabc2ef983467ddd07522256a4cc6634cdc00c8"} Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.582424 4708 scope.go:117] "RemoveContainer" containerID="37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.582742 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.613952 4708 scope.go:117] "RemoveContainer" containerID="0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.622467 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.635664 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.648671 4708 scope.go:117] "RemoveContainer" containerID="37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.651604 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: E1125 05:56:23.652124 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-log" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652138 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-log" Nov 25 05:56:23 crc kubenswrapper[4708]: E1125 05:56:23.652165 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-httpd" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652172 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-httpd" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652336 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-httpd" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652351 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" containerName="glance-log" Nov 25 05:56:23 crc kubenswrapper[4708]: E1125 05:56:23.652694 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e\": container with ID starting with 37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e not found: ID does not exist" containerID="37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652731 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e"} err="failed to get container status \"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e\": rpc error: code = NotFound desc = could not find container \"37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e\": container with ID starting with 37c947a3c7625b7c9df9b0c0d10360e248876dab3f3a38e632e13da5cb83f63e not found: ID does not exist" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.652757 4708 scope.go:117] "RemoveContainer" containerID="0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.653427 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.661107 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.668499 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Nov 25 05:56:23 crc kubenswrapper[4708]: E1125 05:56:23.668511 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7\": container with ID starting with 0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7 not found: ID does not exist" containerID="0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.668559 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7"} err="failed to get container status \"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7\": rpc error: code = NotFound desc = could not find container \"0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7\": container with ID starting with 0582469383e43730839f698fd05c6cbf2aab486fb013da0b5901d8d2a891d1e7 not found: ID does not exist" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.668672 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796533 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fhhj\" (UniqueName: \"kubernetes.io/projected/6891a73d-d8e8-4a12-9136-b56912aeeb54-kube-api-access-5fhhj\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796588 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796741 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796769 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-logs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796813 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796865 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796906 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.796926 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899194 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fhhj\" (UniqueName: \"kubernetes.io/projected/6891a73d-d8e8-4a12-9136-b56912aeeb54-kube-api-access-5fhhj\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899240 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899322 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899340 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-logs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899375 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899420 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899456 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899474 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899831 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-logs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899897 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6891a73d-d8e8-4a12-9136-b56912aeeb54-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.899916 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.907052 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.909190 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.910937 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.913642 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6891a73d-d8e8-4a12-9136-b56912aeeb54-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.924090 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fhhj\" (UniqueName: \"kubernetes.io/projected/6891a73d-d8e8-4a12-9136-b56912aeeb54-kube-api-access-5fhhj\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.929710 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"6891a73d-d8e8-4a12-9136-b56912aeeb54\") " pod="openstack/glance-default-internal-api-0" Nov 25 05:56:23 crc kubenswrapper[4708]: I1125 05:56:23.981475 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.525294 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Nov 25 05:56:24 crc kubenswrapper[4708]: W1125 05:56:24.549129 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6891a73d_d8e8_4a12_9136_b56912aeeb54.slice/crio-bb9af935b3f2f17fad072c55958fd46e61cf97ac178ff30598ccb957f1fcfe17 WatchSource:0}: Error finding container bb9af935b3f2f17fad072c55958fd46e61cf97ac178ff30598ccb957f1fcfe17: Status 404 returned error can't find the container with id bb9af935b3f2f17fad072c55958fd46e61cf97ac178ff30598ccb957f1fcfe17 Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.593195 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0af2b359-a58e-4bc7-b497-40368df40ee1","Type":"ContainerStarted","Data":"c8ff1bba9c2d8e32950ea8161ef7a1c27929acde79266cc7e3ef343c707a92e0"} Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.595383 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6891a73d-d8e8-4a12-9136-b56912aeeb54","Type":"ContainerStarted","Data":"bb9af935b3f2f17fad072c55958fd46e61cf97ac178ff30598ccb957f1fcfe17"} Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.595513 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-central-agent" containerID="cri-o://fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846" gracePeriod=30 Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.595917 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="proxy-httpd" containerID="cri-o://583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987" gracePeriod=30 Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.595979 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-notification-agent" containerID="cri-o://30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369" gracePeriod=30 Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.596058 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="sg-core" containerID="cri-o://0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6" gracePeriod=30 Nov 25 05:56:24 crc kubenswrapper[4708]: I1125 05:56:24.903100 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493b9467-e848-4aae-a5a5-5d699811e40e" path="/var/lib/kubelet/pods/493b9467-e848-4aae-a5a5-5d699811e40e/volumes" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.611774 4708 generic.go:334] "Generic (PLEG): container finished" podID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerID="583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987" exitCode=0 Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.612354 4708 generic.go:334] "Generic (PLEG): container finished" podID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerID="0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6" exitCode=2 Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.612365 4708 generic.go:334] "Generic (PLEG): container finished" podID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerID="30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369" exitCode=0 Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.611870 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerDied","Data":"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.612480 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerDied","Data":"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.612500 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerDied","Data":"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.614860 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0af2b359-a58e-4bc7-b497-40368df40ee1","Type":"ContainerStarted","Data":"b7c45b45f5c8546f41bd69d088848ac2d3e4e9968864d924c2a4321cdc44f1a0"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.617266 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6891a73d-d8e8-4a12-9136-b56912aeeb54","Type":"ContainerStarted","Data":"9519f0d9b2e8d26d5e5682ad8a3b18bd55ef4584f385722029cf5b4dc61a1617"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.617320 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6891a73d-d8e8-4a12-9136-b56912aeeb54","Type":"ContainerStarted","Data":"8d5c3aaf3a4c0f7a41594eb84f19d97f19cd29dce9d946309a3c6ce1de3eb625"} Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.643173 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.643159468 podStartE2EDuration="3.643159468s" podCreationTimestamp="2025-11-25 05:56:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:25.638625754 +0000 UTC m=+927.047459140" watchObservedRunningTime="2025-11-25 05:56:25.643159468 +0000 UTC m=+927.051992854" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.665629 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.665600682 podStartE2EDuration="2.665600682s" podCreationTimestamp="2025-11-25 05:56:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:25.657852469 +0000 UTC m=+927.066685856" watchObservedRunningTime="2025-11-25 05:56:25.665600682 +0000 UTC m=+927.074434058" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.802445 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vhfrh"] Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.803936 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.806583 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.806717 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-d6kz9" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.806792 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.813747 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vhfrh"] Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.844569 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.844627 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w8m2\" (UniqueName: \"kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.844733 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.844757 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.946877 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.946939 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w8m2\" (UniqueName: \"kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.947017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.947037 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.953665 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.954788 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.961560 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:25 crc kubenswrapper[4708]: I1125 05:56:25.963214 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w8m2\" (UniqueName: \"kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2\") pod \"nova-cell0-conductor-db-sync-vhfrh\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:26 crc kubenswrapper[4708]: I1125 05:56:26.117788 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:26 crc kubenswrapper[4708]: I1125 05:56:26.530275 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vhfrh"] Nov 25 05:56:26 crc kubenswrapper[4708]: I1125 05:56:26.635208 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" event={"ID":"7ba92183-cc87-4b0f-8b71-8be8d39be23c","Type":"ContainerStarted","Data":"bc812dd570345e4485b78ed68e10c265cb1659620907520c1931fe8fd49b5cc3"} Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.224326 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312687 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312776 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312816 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312880 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312947 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.312982 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lm22\" (UniqueName: \"kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.313035 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data\") pod \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\" (UID: \"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a\") " Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.313503 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.313948 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.334089 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22" (OuterVolumeSpecName: "kube-api-access-8lm22") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "kube-api-access-8lm22". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.335601 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts" (OuterVolumeSpecName: "scripts") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.344183 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.385167 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.412147 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data" (OuterVolumeSpecName: "config-data") pod "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" (UID: "dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415797 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415863 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415875 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415916 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415930 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415940 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.415951 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lm22\" (UniqueName: \"kubernetes.io/projected/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a-kube-api-access-8lm22\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.710472 4708 generic.go:334] "Generic (PLEG): container finished" podID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerID="fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846" exitCode=0 Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.710546 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerDied","Data":"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846"} Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.710570 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.710587 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a","Type":"ContainerDied","Data":"815602f04cf2347e530f6e595c8d7be7178ca77b9b2e46db1d5f0007ea2cc349"} Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.710611 4708 scope.go:117] "RemoveContainer" containerID="583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.749428 4708 scope.go:117] "RemoveContainer" containerID="0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.757554 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.768740 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782022 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.782585 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="proxy-httpd" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782606 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="proxy-httpd" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.782615 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-central-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782622 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-central-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.782634 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="sg-core" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782641 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="sg-core" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.782657 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-notification-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782662 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-notification-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782856 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-notification-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782874 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="ceilometer-central-agent" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782885 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="proxy-httpd" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.782903 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" containerName="sg-core" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.785215 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.785838 4708 scope.go:117] "RemoveContainer" containerID="30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.788985 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.789072 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.800294 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.814129 4708 scope.go:117] "RemoveContainer" containerID="fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.824971 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825028 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825092 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825127 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqjtb\" (UniqueName: \"kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825151 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825293 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.825342 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.831844 4708 scope.go:117] "RemoveContainer" containerID="583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.832357 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987\": container with ID starting with 583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987 not found: ID does not exist" containerID="583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.832400 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987"} err="failed to get container status \"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987\": rpc error: code = NotFound desc = could not find container \"583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987\": container with ID starting with 583268e8e7e04b17c364e733ce3365aa28376f9f38e317f7f385cb007a3e2987 not found: ID does not exist" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.832435 4708 scope.go:117] "RemoveContainer" containerID="0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.832711 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6\": container with ID starting with 0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6 not found: ID does not exist" containerID="0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.832736 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6"} err="failed to get container status \"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6\": rpc error: code = NotFound desc = could not find container \"0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6\": container with ID starting with 0e316afc0277bf0f0f774d2cfb8fad8acfeeb00810a22cdedb05e8b688e265a6 not found: ID does not exist" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.832748 4708 scope.go:117] "RemoveContainer" containerID="30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.833026 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369\": container with ID starting with 30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369 not found: ID does not exist" containerID="30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.833077 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369"} err="failed to get container status \"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369\": rpc error: code = NotFound desc = could not find container \"30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369\": container with ID starting with 30806ebd2de6178b9426cadb6966ac4815c80b6eac6ceb7527f398132e694369 not found: ID does not exist" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.833111 4708 scope.go:117] "RemoveContainer" containerID="fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846" Nov 25 05:56:32 crc kubenswrapper[4708]: E1125 05:56:32.834033 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846\": container with ID starting with fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846 not found: ID does not exist" containerID="fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.834078 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846"} err="failed to get container status \"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846\": rpc error: code = NotFound desc = could not find container \"fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846\": container with ID starting with fb3d020f30779d3710065e40001693fff1a295b0934125bc85a0dff789579846 not found: ID does not exist" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.908035 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a" path="/var/lib/kubelet/pods/dbad4b9e-87a5-4ae3-9e3b-9a0b6870240a/volumes" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.926732 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.926801 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.927772 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.927851 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.927931 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.927968 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqjtb\" (UniqueName: \"kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.928007 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.928442 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.928633 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.934280 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.934355 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.936336 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.936745 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.945362 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqjtb\" (UniqueName: \"kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb\") pod \"ceilometer-0\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " pod="openstack/ceilometer-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.957876 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.957944 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.987814 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 05:56:32 crc kubenswrapper[4708]: I1125 05:56:32.993877 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Nov 25 05:56:33 crc kubenswrapper[4708]: I1125 05:56:33.101234 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:33 crc kubenswrapper[4708]: I1125 05:56:33.724732 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 05:56:33 crc kubenswrapper[4708]: I1125 05:56:33.724786 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Nov 25 05:56:33 crc kubenswrapper[4708]: I1125 05:56:33.982329 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:33 crc kubenswrapper[4708]: I1125 05:56:33.982962 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:34 crc kubenswrapper[4708]: I1125 05:56:34.010812 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:34 crc kubenswrapper[4708]: I1125 05:56:34.020770 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:34 crc kubenswrapper[4708]: I1125 05:56:34.732595 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:34 crc kubenswrapper[4708]: I1125 05:56:34.732657 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:35 crc kubenswrapper[4708]: I1125 05:56:35.280820 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 05:56:35 crc kubenswrapper[4708]: I1125 05:56:35.281462 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Nov 25 05:56:36 crc kubenswrapper[4708]: I1125 05:56:36.530789 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:36 crc kubenswrapper[4708]: I1125 05:56:36.538656 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Nov 25 05:56:38 crc kubenswrapper[4708]: I1125 05:56:38.724331 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:38 crc kubenswrapper[4708]: I1125 05:56:38.789480 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" event={"ID":"7ba92183-cc87-4b0f-8b71-8be8d39be23c","Type":"ContainerStarted","Data":"bdd00b3b0105ac4475633b45b1144b71cd968505cdb599227905d7c3a349edc1"} Nov 25 05:56:38 crc kubenswrapper[4708]: I1125 05:56:38.790610 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerStarted","Data":"0ae56ee2da52970007146133aab33f4382515b4e45e4cf66fd0ca743e38e6fb7"} Nov 25 05:56:38 crc kubenswrapper[4708]: I1125 05:56:38.801249 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" podStartSLOduration=1.984692142 podStartE2EDuration="13.801240108s" podCreationTimestamp="2025-11-25 05:56:25 +0000 UTC" firstStartedPulling="2025-11-25 05:56:26.540109487 +0000 UTC m=+927.948942873" lastFinishedPulling="2025-11-25 05:56:38.356657463 +0000 UTC m=+939.765490839" observedRunningTime="2025-11-25 05:56:38.800453515 +0000 UTC m=+940.209286901" watchObservedRunningTime="2025-11-25 05:56:38.801240108 +0000 UTC m=+940.210073493" Nov 25 05:56:39 crc kubenswrapper[4708]: I1125 05:56:39.801365 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerStarted","Data":"23617bc9065a576531abbeaaf5adeb8d4820676ac3d5ac7325441329e2022e67"} Nov 25 05:56:40 crc kubenswrapper[4708]: I1125 05:56:40.813258 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerStarted","Data":"f210033a23a7659992def92ac2e862fff557861d2d235c5e6d3e201625be16c3"} Nov 25 05:56:41 crc kubenswrapper[4708]: I1125 05:56:41.846547 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerStarted","Data":"e7726ef45cb0248c00ddbd1e63460e2523fc6d48387cc4da934390db2729ed39"} Nov 25 05:56:43 crc kubenswrapper[4708]: I1125 05:56:43.869507 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerStarted","Data":"6643c2b6176a2d6023979254e681f1adf19a47f4747abb4b84e0a5b8b68645de"} Nov 25 05:56:43 crc kubenswrapper[4708]: I1125 05:56:43.871184 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:56:43 crc kubenswrapper[4708]: I1125 05:56:43.906595 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.031181682 podStartE2EDuration="11.906580194s" podCreationTimestamp="2025-11-25 05:56:32 +0000 UTC" firstStartedPulling="2025-11-25 05:56:38.727637491 +0000 UTC m=+940.136470878" lastFinishedPulling="2025-11-25 05:56:42.603036003 +0000 UTC m=+944.011869390" observedRunningTime="2025-11-25 05:56:43.901978501 +0000 UTC m=+945.310811887" watchObservedRunningTime="2025-11-25 05:56:43.906580194 +0000 UTC m=+945.315413569" Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.356790 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.356856 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.356908 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.357805 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.357867 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e" gracePeriod=600 Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.881638 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e" exitCode=0 Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.881691 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e"} Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.881932 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307"} Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.881958 4708 scope.go:117] "RemoveContainer" containerID="c6b477dd423864905be00ad5487d4b0770af74605bb3452dda5b50dab978b977" Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.886798 4708 generic.go:334] "Generic (PLEG): container finished" podID="7ba92183-cc87-4b0f-8b71-8be8d39be23c" containerID="bdd00b3b0105ac4475633b45b1144b71cd968505cdb599227905d7c3a349edc1" exitCode=0 Nov 25 05:56:44 crc kubenswrapper[4708]: I1125 05:56:44.887842 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" event={"ID":"7ba92183-cc87-4b0f-8b71-8be8d39be23c","Type":"ContainerDied","Data":"bdd00b3b0105ac4475633b45b1144b71cd968505cdb599227905d7c3a349edc1"} Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.241607 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.434334 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle\") pod \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.434440 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data\") pod \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.434468 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts\") pod \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.434572 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w8m2\" (UniqueName: \"kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2\") pod \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\" (UID: \"7ba92183-cc87-4b0f-8b71-8be8d39be23c\") " Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.447698 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts" (OuterVolumeSpecName: "scripts") pod "7ba92183-cc87-4b0f-8b71-8be8d39be23c" (UID: "7ba92183-cc87-4b0f-8b71-8be8d39be23c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.447750 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2" (OuterVolumeSpecName: "kube-api-access-2w8m2") pod "7ba92183-cc87-4b0f-8b71-8be8d39be23c" (UID: "7ba92183-cc87-4b0f-8b71-8be8d39be23c"). InnerVolumeSpecName "kube-api-access-2w8m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.461918 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data" (OuterVolumeSpecName: "config-data") pod "7ba92183-cc87-4b0f-8b71-8be8d39be23c" (UID: "7ba92183-cc87-4b0f-8b71-8be8d39be23c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.464288 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ba92183-cc87-4b0f-8b71-8be8d39be23c" (UID: "7ba92183-cc87-4b0f-8b71-8be8d39be23c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.538755 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.538792 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.538804 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ba92183-cc87-4b0f-8b71-8be8d39be23c-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.538815 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w8m2\" (UniqueName: \"kubernetes.io/projected/7ba92183-cc87-4b0f-8b71-8be8d39be23c-kube-api-access-2w8m2\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.915331 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" event={"ID":"7ba92183-cc87-4b0f-8b71-8be8d39be23c","Type":"ContainerDied","Data":"bc812dd570345e4485b78ed68e10c265cb1659620907520c1931fe8fd49b5cc3"} Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.915694 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc812dd570345e4485b78ed68e10c265cb1659620907520c1931fe8fd49b5cc3" Nov 25 05:56:46 crc kubenswrapper[4708]: I1125 05:56:46.915802 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-vhfrh" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.006845 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 05:56:47 crc kubenswrapper[4708]: E1125 05:56:47.007284 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ba92183-cc87-4b0f-8b71-8be8d39be23c" containerName="nova-cell0-conductor-db-sync" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.007304 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ba92183-cc87-4b0f-8b71-8be8d39be23c" containerName="nova-cell0-conductor-db-sync" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.007463 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ba92183-cc87-4b0f-8b71-8be8d39be23c" containerName="nova-cell0-conductor-db-sync" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.008128 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.009986 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-d6kz9" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.010037 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.017317 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.152560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.152633 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.152684 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx2nv\" (UniqueName: \"kubernetes.io/projected/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-kube-api-access-xx2nv\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.255515 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.255664 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.255799 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx2nv\" (UniqueName: \"kubernetes.io/projected/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-kube-api-access-xx2nv\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.260854 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.260906 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.271367 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx2nv\" (UniqueName: \"kubernetes.io/projected/f6d2e9ec-0ff0-4db9-af17-51d9e824252d-kube-api-access-xx2nv\") pod \"nova-cell0-conductor-0\" (UID: \"f6d2e9ec-0ff0-4db9-af17-51d9e824252d\") " pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.321686 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.738697 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.924786 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f6d2e9ec-0ff0-4db9-af17-51d9e824252d","Type":"ContainerStarted","Data":"a06787f11090086c72e821870379073f781684739aac2ed39ca0f5412153ae14"} Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.925120 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"f6d2e9ec-0ff0-4db9-af17-51d9e824252d","Type":"ContainerStarted","Data":"f4177f72bdcb56d2183fe61c31e0dbacd58b8371ed0641dc7f2722ec996a96b4"} Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.925891 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:47 crc kubenswrapper[4708]: I1125 05:56:47.943639 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.943622489 podStartE2EDuration="1.943622489s" podCreationTimestamp="2025-11-25 05:56:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:47.936846778 +0000 UTC m=+949.345680165" watchObservedRunningTime="2025-11-25 05:56:47.943622489 +0000 UTC m=+949.352455864" Nov 25 05:56:51 crc kubenswrapper[4708]: I1125 05:56:51.856816 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:51 crc kubenswrapper[4708]: I1125 05:56:51.857327 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-central-agent" containerID="cri-o://23617bc9065a576531abbeaaf5adeb8d4820676ac3d5ac7325441329e2022e67" gracePeriod=30 Nov 25 05:56:51 crc kubenswrapper[4708]: I1125 05:56:51.857380 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="proxy-httpd" containerID="cri-o://6643c2b6176a2d6023979254e681f1adf19a47f4747abb4b84e0a5b8b68645de" gracePeriod=30 Nov 25 05:56:51 crc kubenswrapper[4708]: I1125 05:56:51.857403 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-notification-agent" containerID="cri-o://f210033a23a7659992def92ac2e862fff557861d2d235c5e6d3e201625be16c3" gracePeriod=30 Nov 25 05:56:51 crc kubenswrapper[4708]: I1125 05:56:51.857380 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="sg-core" containerID="cri-o://e7726ef45cb0248c00ddbd1e63460e2523fc6d48387cc4da934390db2729ed39" gracePeriod=30 Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.346542 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.808139 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-5gq69"] Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.809566 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.812661 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.812725 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.816744 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5gq69"] Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.863513 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.863608 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlg48\" (UniqueName: \"kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.863660 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.864021 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.947653 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.948867 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.951038 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.963675 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.966099 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.966172 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.966223 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.968137 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlg48\" (UniqueName: \"kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.968223 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqlh8\" (UniqueName: \"kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.968294 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.968424 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.981026 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.997232 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.997256 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:52 crc kubenswrapper[4708]: I1125 05:56:52.997736 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlg48\" (UniqueName: \"kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48\") pod \"nova-cell0-cell-mapping-5gq69\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022736 4708 generic.go:334] "Generic (PLEG): container finished" podID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerID="6643c2b6176a2d6023979254e681f1adf19a47f4747abb4b84e0a5b8b68645de" exitCode=0 Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022786 4708 generic.go:334] "Generic (PLEG): container finished" podID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerID="e7726ef45cb0248c00ddbd1e63460e2523fc6d48387cc4da934390db2729ed39" exitCode=2 Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022793 4708 generic.go:334] "Generic (PLEG): container finished" podID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerID="23617bc9065a576531abbeaaf5adeb8d4820676ac3d5ac7325441329e2022e67" exitCode=0 Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022820 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerDied","Data":"6643c2b6176a2d6023979254e681f1adf19a47f4747abb4b84e0a5b8b68645de"} Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022852 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerDied","Data":"e7726ef45cb0248c00ddbd1e63460e2523fc6d48387cc4da934390db2729ed39"} Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.022864 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerDied","Data":"23617bc9065a576531abbeaaf5adeb8d4820676ac3d5ac7325441329e2022e67"} Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.087045 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqlh8\" (UniqueName: \"kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.087202 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.087400 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.090732 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.093838 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.104374 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.141204 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.144570 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.151029 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.151471 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.164008 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqlh8\" (UniqueName: \"kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8\") pod \"nova-scheduler-0\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.193239 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.208799 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.213818 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.216623 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.219239 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.219281 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbjn8\" (UniqueName: \"kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.219354 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.219478 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.231594 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.233816 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.236485 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.238753 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.241712 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.243343 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.246589 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.266370 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321449 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321500 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321542 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321565 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321583 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbjn8\" (UniqueName: \"kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321599 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321623 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321644 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321662 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321679 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321696 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321717 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4clb\" (UniqueName: \"kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321744 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321769 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxfxl\" (UniqueName: \"kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321812 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tws7\" (UniqueName: \"kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321832 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.321847 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.322217 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.328737 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.328779 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.338147 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbjn8\" (UniqueName: \"kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8\") pod \"nova-metadata-0\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.423648 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.423888 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.423918 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.423945 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424004 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4clb\" (UniqueName: \"kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424121 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxfxl\" (UniqueName: \"kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424218 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tws7\" (UniqueName: \"kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424250 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424282 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424367 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424404 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424446 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.424483 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.425657 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.425694 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.426867 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.427810 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.428771 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.430411 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.431419 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.434073 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.438742 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4clb\" (UniqueName: \"kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.438765 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.443175 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxfxl\" (UniqueName: \"kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl\") pod \"nova-cell1-novncproxy-0\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.444789 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.449845 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tws7\" (UniqueName: \"kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7\") pod \"dnsmasq-dns-d6558c9-hghrr\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.532548 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.544971 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.565836 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.578751 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.733612 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-5gq69"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.810297 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:56:53 crc kubenswrapper[4708]: W1125 05:56:53.813054 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f16b36f_cb7d_4820_808c_e3fccd217e90.slice/crio-152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0 WatchSource:0}: Error finding container 152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0: Status 404 returned error can't find the container with id 152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0 Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.971661 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbbt5"] Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.972974 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.976352 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.976480 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 05:56:53 crc kubenswrapper[4708]: I1125 05:56:53.978691 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbbt5"] Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.030362 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:56:54 crc kubenswrapper[4708]: W1125 05:56:54.035322 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee4549fd_ebcf_4169_b8cf_04ab1e995f9f.slice/crio-34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925 WatchSource:0}: Error finding container 34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925: Status 404 returned error can't find the container with id 34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925 Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.035385 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5gq69" event={"ID":"527c5c0e-4e68-4533-b9b0-e65cfd465ce7","Type":"ContainerStarted","Data":"b7130d42a3fdc3f2702b6ea2382f13f367557e7becf529cc8e8a14f39882f4a4"} Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.035421 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5gq69" event={"ID":"527c5c0e-4e68-4533-b9b0-e65cfd465ce7","Type":"ContainerStarted","Data":"ab011b7dd4147eef07d453607a8c584a672d5e4d861c5add16a96d9811bb145b"} Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.039249 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f16b36f-cb7d-4820-808c-e3fccd217e90","Type":"ContainerStarted","Data":"152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0"} Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.041062 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlncr\" (UniqueName: \"kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.041276 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.041382 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.041683 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.057905 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-5gq69" podStartSLOduration=2.05788397 podStartE2EDuration="2.05788397s" podCreationTimestamp="2025-11-25 05:56:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:54.050063031 +0000 UTC m=+955.458896417" watchObservedRunningTime="2025-11-25 05:56:54.05788397 +0000 UTC m=+955.466717357" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.112938 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.144075 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.144595 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlncr\" (UniqueName: \"kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.144771 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.145512 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.149372 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.150669 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.151202 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.160579 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlncr\" (UniqueName: \"kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr\") pod \"nova-cell1-conductor-db-sync-vbbt5\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.196723 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.218494 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:56:54 crc kubenswrapper[4708]: W1125 05:56:54.218854 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6c52ba9_bea9_4756_9c31_f2a374c71674.slice/crio-58be7982844aa6d26a7b1b7f3a8501ebe0e6ef02c0f47d331306c49f74b331a9 WatchSource:0}: Error finding container 58be7982844aa6d26a7b1b7f3a8501ebe0e6ef02c0f47d331306c49f74b331a9: Status 404 returned error can't find the container with id 58be7982844aa6d26a7b1b7f3a8501ebe0e6ef02c0f47d331306c49f74b331a9 Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.381808 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:56:54 crc kubenswrapper[4708]: I1125 05:56:54.705849 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbbt5"] Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.091981 4708 generic.go:334] "Generic (PLEG): container finished" podID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerID="e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592" exitCode=0 Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.092318 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6558c9-hghrr" event={"ID":"d954cc45-5468-4663-ae00-9076e5ecae8e","Type":"ContainerDied","Data":"e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.092704 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6558c9-hghrr" event={"ID":"d954cc45-5468-4663-ae00-9076e5ecae8e","Type":"ContainerStarted","Data":"15e4ee1fa7f63d764b4c2bb728cebd995b32a4749d0976f9ae9f8846f68e222f"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.099976 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerStarted","Data":"58f2c299726dbbbfa7d4579039c9782ca7bde75f3ce1fafae2bc1440421246e4"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.127449 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerStarted","Data":"34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.136020 4708 generic.go:334] "Generic (PLEG): container finished" podID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerID="f210033a23a7659992def92ac2e862fff557861d2d235c5e6d3e201625be16c3" exitCode=0 Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.136215 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerDied","Data":"f210033a23a7659992def92ac2e862fff557861d2d235c5e6d3e201625be16c3"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.139542 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6c52ba9-bea9-4756-9c31-f2a374c71674","Type":"ContainerStarted","Data":"58be7982844aa6d26a7b1b7f3a8501ebe0e6ef02c0f47d331306c49f74b331a9"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.147814 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" event={"ID":"f2654acf-c93c-4c8d-bd73-537b9c828615","Type":"ContainerStarted","Data":"9ae42d476d5f2424caf80504e7ddcc8023c2bc5417abd3fb858475768879d891"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.151677 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" event={"ID":"f2654acf-c93c-4c8d-bd73-537b9c828615","Type":"ContainerStarted","Data":"f6d3116bc56327f2e019d7528f587a15d1d51880454264297a903f8de526e4d3"} Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.165031 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" podStartSLOduration=2.165011492 podStartE2EDuration="2.165011492s" podCreationTimestamp="2025-11-25 05:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:55.163950523 +0000 UTC m=+956.572783909" watchObservedRunningTime="2025-11-25 05:56:55.165011492 +0000 UTC m=+956.573844878" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.432586 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512395 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512741 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512789 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512904 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqjtb\" (UniqueName: \"kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512946 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512959 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.512991 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle\") pod \"1bc1700f-cae5-43fc-8c66-581bbf53d321\" (UID: \"1bc1700f-cae5-43fc-8c66-581bbf53d321\") " Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.513561 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.513708 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.514041 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.514059 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1bc1700f-cae5-43fc-8c66-581bbf53d321-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.519236 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts" (OuterVolumeSpecName: "scripts") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.519664 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb" (OuterVolumeSpecName: "kube-api-access-gqjtb") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "kube-api-access-gqjtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.553643 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.605080 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.619875 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqjtb\" (UniqueName: \"kubernetes.io/projected/1bc1700f-cae5-43fc-8c66-581bbf53d321-kube-api-access-gqjtb\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.619917 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.619927 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.619937 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.631582 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data" (OuterVolumeSpecName: "config-data") pod "1bc1700f-cae5-43fc-8c66-581bbf53d321" (UID: "1bc1700f-cae5-43fc-8c66-581bbf53d321"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:55 crc kubenswrapper[4708]: I1125 05:56:55.722452 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1bc1700f-cae5-43fc-8c66-581bbf53d321-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.183009 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1bc1700f-cae5-43fc-8c66-581bbf53d321","Type":"ContainerDied","Data":"0ae56ee2da52970007146133aab33f4382515b4e45e4cf66fd0ca743e38e6fb7"} Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.183061 4708 scope.go:117] "RemoveContainer" containerID="6643c2b6176a2d6023979254e681f1adf19a47f4747abb4b84e0a5b8b68645de" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.183175 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.189591 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6558c9-hghrr" event={"ID":"d954cc45-5468-4663-ae00-9076e5ecae8e","Type":"ContainerStarted","Data":"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448"} Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.189775 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.208330 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-d6558c9-hghrr" podStartSLOduration=3.208319827 podStartE2EDuration="3.208319827s" podCreationTimestamp="2025-11-25 05:56:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:56:56.205048591 +0000 UTC m=+957.613881978" watchObservedRunningTime="2025-11-25 05:56:56.208319827 +0000 UTC m=+957.617153214" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.238172 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.252629 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262035 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: E1125 05:56:56.262531 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-central-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262553 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-central-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: E1125 05:56:56.262604 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="proxy-httpd" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262610 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="proxy-httpd" Nov 25 05:56:56 crc kubenswrapper[4708]: E1125 05:56:56.262620 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="sg-core" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262626 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="sg-core" Nov 25 05:56:56 crc kubenswrapper[4708]: E1125 05:56:56.262658 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-notification-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262666 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-notification-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262924 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-notification-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262946 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="ceilometer-central-agent" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262956 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="sg-core" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.262973 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" containerName="proxy-httpd" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.264849 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.266791 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.267078 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.273364 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337011 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337067 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337099 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337143 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337176 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337197 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.337212 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkz72\" (UniqueName: \"kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.342912 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.352993 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441166 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441214 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441237 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441267 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441291 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441312 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.441326 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkz72\" (UniqueName: \"kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.442025 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.442827 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.446383 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.447954 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.449692 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.455942 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.456215 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkz72\" (UniqueName: \"kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72\") pod \"ceilometer-0\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.595616 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:56:56 crc kubenswrapper[4708]: I1125 05:56:56.909637 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bc1700f-cae5-43fc-8c66-581bbf53d321" path="/var/lib/kubelet/pods/1bc1700f-cae5-43fc-8c66-581bbf53d321/volumes" Nov 25 05:56:57 crc kubenswrapper[4708]: I1125 05:56:57.285453 4708 scope.go:117] "RemoveContainer" containerID="e7726ef45cb0248c00ddbd1e63460e2523fc6d48387cc4da934390db2729ed39" Nov 25 05:56:57 crc kubenswrapper[4708]: I1125 05:56:57.389206 4708 scope.go:117] "RemoveContainer" containerID="f210033a23a7659992def92ac2e862fff557861d2d235c5e6d3e201625be16c3" Nov 25 05:56:57 crc kubenswrapper[4708]: I1125 05:56:57.447625 4708 scope.go:117] "RemoveContainer" containerID="23617bc9065a576531abbeaaf5adeb8d4820676ac3d5ac7325441329e2022e67" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.004844 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.212479 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f16b36f-cb7d-4820-808c-e3fccd217e90","Type":"ContainerStarted","Data":"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.213964 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerStarted","Data":"86120dae20bc46177936695c106a0a42139200feacfc2756fa01d97c05b8e617"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.216278 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerStarted","Data":"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.216386 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerStarted","Data":"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.218813 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerStarted","Data":"d1b62f640775a62574f016487b8923708155f6b4faec0f89ab88ffedb84fa469"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.218897 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerStarted","Data":"f4545565f0ba9f494ea8dbabb1f37fc4ee11a43cb1cb0effec87cc232686469c"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.219050 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-log" containerID="cri-o://f4545565f0ba9f494ea8dbabb1f37fc4ee11a43cb1cb0effec87cc232686469c" gracePeriod=30 Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.219322 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-metadata" containerID="cri-o://d1b62f640775a62574f016487b8923708155f6b4faec0f89ab88ffedb84fa469" gracePeriod=30 Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.225598 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6c52ba9-bea9-4756-9c31-f2a374c71674","Type":"ContainerStarted","Data":"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181"} Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.225747 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="d6c52ba9-bea9-4756-9c31-f2a374c71674" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181" gracePeriod=30 Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.230103 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.651554895 podStartE2EDuration="6.230092106s" podCreationTimestamp="2025-11-25 05:56:52 +0000 UTC" firstStartedPulling="2025-11-25 05:56:53.816121462 +0000 UTC m=+955.224954849" lastFinishedPulling="2025-11-25 05:56:57.394658674 +0000 UTC m=+958.803492060" observedRunningTime="2025-11-25 05:56:58.228433651 +0000 UTC m=+959.637267037" watchObservedRunningTime="2025-11-25 05:56:58.230092106 +0000 UTC m=+959.638925493" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.266713 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.274844 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.924430091 podStartE2EDuration="5.274826386s" podCreationTimestamp="2025-11-25 05:56:53 +0000 UTC" firstStartedPulling="2025-11-25 05:56:54.038787992 +0000 UTC m=+955.447621378" lastFinishedPulling="2025-11-25 05:56:57.389184287 +0000 UTC m=+958.798017673" observedRunningTime="2025-11-25 05:56:58.26105976 +0000 UTC m=+959.669893145" watchObservedRunningTime="2025-11-25 05:56:58.274826386 +0000 UTC m=+959.683659771" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.284667 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.091433474 podStartE2EDuration="5.284651302s" podCreationTimestamp="2025-11-25 05:56:53 +0000 UTC" firstStartedPulling="2025-11-25 05:56:54.228608117 +0000 UTC m=+955.637441503" lastFinishedPulling="2025-11-25 05:56:57.421825945 +0000 UTC m=+958.830659331" observedRunningTime="2025-11-25 05:56:58.274166422 +0000 UTC m=+959.682999808" watchObservedRunningTime="2025-11-25 05:56:58.284651302 +0000 UTC m=+959.693484688" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.297560 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.020614123 podStartE2EDuration="5.297509435s" podCreationTimestamp="2025-11-25 05:56:53 +0000 UTC" firstStartedPulling="2025-11-25 05:56:54.112352846 +0000 UTC m=+955.521186231" lastFinishedPulling="2025-11-25 05:56:57.389248157 +0000 UTC m=+958.798081543" observedRunningTime="2025-11-25 05:56:58.289370406 +0000 UTC m=+959.698203792" watchObservedRunningTime="2025-11-25 05:56:58.297509435 +0000 UTC m=+959.706342821" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.533176 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.533597 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:56:58 crc kubenswrapper[4708]: I1125 05:56:58.580002 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.243579 4708 generic.go:334] "Generic (PLEG): container finished" podID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerID="d1b62f640775a62574f016487b8923708155f6b4faec0f89ab88ffedb84fa469" exitCode=0 Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.243618 4708 generic.go:334] "Generic (PLEG): container finished" podID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerID="f4545565f0ba9f494ea8dbabb1f37fc4ee11a43cb1cb0effec87cc232686469c" exitCode=143 Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.244616 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerDied","Data":"d1b62f640775a62574f016487b8923708155f6b4faec0f89ab88ffedb84fa469"} Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.244676 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerDied","Data":"f4545565f0ba9f494ea8dbabb1f37fc4ee11a43cb1cb0effec87cc232686469c"} Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.542638 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.615979 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs\") pod \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.616058 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data\") pod \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.616146 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbjn8\" (UniqueName: \"kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8\") pod \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.616283 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle\") pod \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\" (UID: \"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f\") " Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.617225 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs" (OuterVolumeSpecName: "logs") pod "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" (UID: "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.621939 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8" (OuterVolumeSpecName: "kube-api-access-nbjn8") pod "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" (UID: "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f"). InnerVolumeSpecName "kube-api-access-nbjn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.646584 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" (UID: "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.649616 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data" (OuterVolumeSpecName: "config-data") pod "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" (UID: "ee4549fd-ebcf-4169-b8cf-04ab1e995f9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.718767 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.718805 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.718818 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:56:59 crc kubenswrapper[4708]: I1125 05:56:59.718842 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbjn8\" (UniqueName: \"kubernetes.io/projected/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f-kube-api-access-nbjn8\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.253683 4708 generic.go:334] "Generic (PLEG): container finished" podID="f2654acf-c93c-4c8d-bd73-537b9c828615" containerID="9ae42d476d5f2424caf80504e7ddcc8023c2bc5417abd3fb858475768879d891" exitCode=0 Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.253964 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" event={"ID":"f2654acf-c93c-4c8d-bd73-537b9c828615","Type":"ContainerDied","Data":"9ae42d476d5f2424caf80504e7ddcc8023c2bc5417abd3fb858475768879d891"} Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.257595 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerStarted","Data":"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e"} Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.257647 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerStarted","Data":"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df"} Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.259768 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.259855 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ee4549fd-ebcf-4169-b8cf-04ab1e995f9f","Type":"ContainerDied","Data":"34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925"} Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.259931 4708 scope.go:117] "RemoveContainer" containerID="d1b62f640775a62574f016487b8923708155f6b4faec0f89ab88ffedb84fa469" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.293580 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.307666 4708 scope.go:117] "RemoveContainer" containerID="f4545565f0ba9f494ea8dbabb1f37fc4ee11a43cb1cb0effec87cc232686469c" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.311495 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.336101 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:00 crc kubenswrapper[4708]: E1125 05:57:00.336552 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-log" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.336571 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-log" Nov 25 05:57:00 crc kubenswrapper[4708]: E1125 05:57:00.336602 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-metadata" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.336610 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-metadata" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.336823 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-metadata" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.336857 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" containerName="nova-metadata-log" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.337879 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.342221 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.343463 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.349473 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.435390 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.435509 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.435580 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.435668 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.435723 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w7jr\" (UniqueName: \"kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.537486 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.537602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w7jr\" (UniqueName: \"kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.537679 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.537779 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.537815 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.538381 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.544649 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.544890 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.544965 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.565263 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w7jr\" (UniqueName: \"kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr\") pod \"nova-metadata-0\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.680543 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:00 crc kubenswrapper[4708]: I1125 05:57:00.906413 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee4549fd-ebcf-4169-b8cf-04ab1e995f9f" path="/var/lib/kubelet/pods/ee4549fd-ebcf-4169-b8cf-04ab1e995f9f/volumes" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.151583 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.272261 4708 generic.go:334] "Generic (PLEG): container finished" podID="527c5c0e-4e68-4533-b9b0-e65cfd465ce7" containerID="b7130d42a3fdc3f2702b6ea2382f13f367557e7becf529cc8e8a14f39882f4a4" exitCode=0 Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.272344 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5gq69" event={"ID":"527c5c0e-4e68-4533-b9b0-e65cfd465ce7","Type":"ContainerDied","Data":"b7130d42a3fdc3f2702b6ea2382f13f367557e7becf529cc8e8a14f39882f4a4"} Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.279341 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerStarted","Data":"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7"} Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.281565 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerStarted","Data":"90bd4268717525a33e49757207c8b1745b42436c52320aba36521e1446a24b79"} Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.606710 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.694537 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data\") pod \"f2654acf-c93c-4c8d-bd73-537b9c828615\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.694923 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle\") pod \"f2654acf-c93c-4c8d-bd73-537b9c828615\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.694989 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlncr\" (UniqueName: \"kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr\") pod \"f2654acf-c93c-4c8d-bd73-537b9c828615\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.695051 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts\") pod \"f2654acf-c93c-4c8d-bd73-537b9c828615\" (UID: \"f2654acf-c93c-4c8d-bd73-537b9c828615\") " Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.710763 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts" (OuterVolumeSpecName: "scripts") pod "f2654acf-c93c-4c8d-bd73-537b9c828615" (UID: "f2654acf-c93c-4c8d-bd73-537b9c828615"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.712707 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr" (OuterVolumeSpecName: "kube-api-access-zlncr") pod "f2654acf-c93c-4c8d-bd73-537b9c828615" (UID: "f2654acf-c93c-4c8d-bd73-537b9c828615"). InnerVolumeSpecName "kube-api-access-zlncr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.724208 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2654acf-c93c-4c8d-bd73-537b9c828615" (UID: "f2654acf-c93c-4c8d-bd73-537b9c828615"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.725407 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data" (OuterVolumeSpecName: "config-data") pod "f2654acf-c93c-4c8d-bd73-537b9c828615" (UID: "f2654acf-c93c-4c8d-bd73-537b9c828615"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.797770 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlncr\" (UniqueName: \"kubernetes.io/projected/f2654acf-c93c-4c8d-bd73-537b9c828615-kube-api-access-zlncr\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.797817 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.797829 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:01 crc kubenswrapper[4708]: I1125 05:57:01.797840 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2654acf-c93c-4c8d-bd73-537b9c828615-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.298754 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerStarted","Data":"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5"} Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.299064 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerStarted","Data":"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc"} Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.302284 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" event={"ID":"f2654acf-c93c-4c8d-bd73-537b9c828615","Type":"ContainerDied","Data":"f6d3116bc56327f2e019d7528f587a15d1d51880454264297a903f8de526e4d3"} Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.302318 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vbbt5" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.302333 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6d3116bc56327f2e019d7528f587a15d1d51880454264297a903f8de526e4d3" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.346027 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.34600761 podStartE2EDuration="2.34600761s" podCreationTimestamp="2025-11-25 05:57:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:02.334456831 +0000 UTC m=+963.743290217" watchObservedRunningTime="2025-11-25 05:57:02.34600761 +0000 UTC m=+963.754840996" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.375674 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 05:57:02 crc kubenswrapper[4708]: E1125 05:57:02.376280 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2654acf-c93c-4c8d-bd73-537b9c828615" containerName="nova-cell1-conductor-db-sync" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.376458 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2654acf-c93c-4c8d-bd73-537b9c828615" containerName="nova-cell1-conductor-db-sync" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.377289 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2654acf-c93c-4c8d-bd73-537b9c828615" containerName="nova-cell1-conductor-db-sync" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.378076 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.379834 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.401057 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.515206 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.515647 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq74f\" (UniqueName: \"kubernetes.io/projected/6b0f9369-64c1-47a1-83ca-052238803461-kube-api-access-zq74f\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.515735 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.619182 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.619642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq74f\" (UniqueName: \"kubernetes.io/projected/6b0f9369-64c1-47a1-83ca-052238803461-kube-api-access-zq74f\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.619786 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.624258 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.625266 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0f9369-64c1-47a1-83ca-052238803461-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.636145 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq74f\" (UniqueName: \"kubernetes.io/projected/6b0f9369-64c1-47a1-83ca-052238803461-kube-api-access-zq74f\") pod \"nova-cell1-conductor-0\" (UID: \"6b0f9369-64c1-47a1-83ca-052238803461\") " pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.648649 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.709408 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.720584 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlg48\" (UniqueName: \"kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48\") pod \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.720752 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data\") pod \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.721141 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts\") pod \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.721211 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle\") pod \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\" (UID: \"527c5c0e-4e68-4533-b9b0-e65cfd465ce7\") " Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.727231 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts" (OuterVolumeSpecName: "scripts") pod "527c5c0e-4e68-4533-b9b0-e65cfd465ce7" (UID: "527c5c0e-4e68-4533-b9b0-e65cfd465ce7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.727384 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48" (OuterVolumeSpecName: "kube-api-access-jlg48") pod "527c5c0e-4e68-4533-b9b0-e65cfd465ce7" (UID: "527c5c0e-4e68-4533-b9b0-e65cfd465ce7"). InnerVolumeSpecName "kube-api-access-jlg48". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.749324 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "527c5c0e-4e68-4533-b9b0-e65cfd465ce7" (UID: "527c5c0e-4e68-4533-b9b0-e65cfd465ce7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.749990 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data" (OuterVolumeSpecName: "config-data") pod "527c5c0e-4e68-4533-b9b0-e65cfd465ce7" (UID: "527c5c0e-4e68-4533-b9b0-e65cfd465ce7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.823838 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.824078 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.824090 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlg48\" (UniqueName: \"kubernetes.io/projected/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-kube-api-access-jlg48\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:02 crc kubenswrapper[4708]: I1125 05:57:02.824100 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/527c5c0e-4e68-4533-b9b0-e65cfd465ce7-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.126667 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.267541 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.294770 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.327580 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerStarted","Data":"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e"} Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.327868 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.329289 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6b0f9369-64c1-47a1-83ca-052238803461","Type":"ContainerStarted","Data":"be6f0006e2c0f9c9650917bd6d12bdd82d928f0a0b9031199e75908887bdbf10"} Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.329420 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"6b0f9369-64c1-47a1-83ca-052238803461","Type":"ContainerStarted","Data":"19a13568da89940e11f8e624c709cfbb75007539fcea54510d65e577c32b402d"} Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.329496 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.331917 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-5gq69" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.331965 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-5gq69" event={"ID":"527c5c0e-4e68-4533-b9b0-e65cfd465ce7","Type":"ContainerDied","Data":"ab011b7dd4147eef07d453607a8c584a672d5e4d861c5add16a96d9811bb145b"} Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.333002 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab011b7dd4147eef07d453607a8c584a672d5e4d861c5add16a96d9811bb145b" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.352906 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.907797666 podStartE2EDuration="7.352889895s" podCreationTimestamp="2025-11-25 05:56:56 +0000 UTC" firstStartedPulling="2025-11-25 05:56:58.014723382 +0000 UTC m=+959.423556767" lastFinishedPulling="2025-11-25 05:57:02.459815609 +0000 UTC m=+963.868648996" observedRunningTime="2025-11-25 05:57:03.344460918 +0000 UTC m=+964.753294304" watchObservedRunningTime="2025-11-25 05:57:03.352889895 +0000 UTC m=+964.761723280" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.366154 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.366132431 podStartE2EDuration="1.366132431s" podCreationTimestamp="2025-11-25 05:57:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:03.358826303 +0000 UTC m=+964.767659689" watchObservedRunningTime="2025-11-25 05:57:03.366132431 +0000 UTC m=+964.774965817" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.369227 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.464351 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.464638 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-log" containerID="cri-o://ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" gracePeriod=30 Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.464721 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-api" containerID="cri-o://5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" gracePeriod=30 Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.480575 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.567897 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.677561 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.679241 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="dnsmasq-dns" containerID="cri-o://d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56" gracePeriod=10 Nov 25 05:57:03 crc kubenswrapper[4708]: I1125 05:57:03.959176 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.077760 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.151589 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.159570 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle\") pod \"778cae07-e592-49fd-ac10-02366a5886a6\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.159613 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs\") pod \"778cae07-e592-49fd-ac10-02366a5886a6\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.159876 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4clb\" (UniqueName: \"kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb\") pod \"778cae07-e592-49fd-ac10-02366a5886a6\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.159951 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data\") pod \"778cae07-e592-49fd-ac10-02366a5886a6\" (UID: \"778cae07-e592-49fd-ac10-02366a5886a6\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.161049 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs" (OuterVolumeSpecName: "logs") pod "778cae07-e592-49fd-ac10-02366a5886a6" (UID: "778cae07-e592-49fd-ac10-02366a5886a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.170674 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb" (OuterVolumeSpecName: "kube-api-access-w4clb") pod "778cae07-e592-49fd-ac10-02366a5886a6" (UID: "778cae07-e592-49fd-ac10-02366a5886a6"). InnerVolumeSpecName "kube-api-access-w4clb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.201576 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "778cae07-e592-49fd-ac10-02366a5886a6" (UID: "778cae07-e592-49fd-ac10-02366a5886a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.211503 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data" (OuterVolumeSpecName: "config-data") pod "778cae07-e592-49fd-ac10-02366a5886a6" (UID: "778cae07-e592-49fd-ac10-02366a5886a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.262487 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.262698 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.262764 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.262827 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.262897 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lx9ns\" (UniqueName: \"kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.263157 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config\") pod \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\" (UID: \"389c1d0d-ddc1-4113-9bdb-2142af81e18f\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.265050 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4clb\" (UniqueName: \"kubernetes.io/projected/778cae07-e592-49fd-ac10-02366a5886a6-kube-api-access-w4clb\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.265136 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.265208 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/778cae07-e592-49fd-ac10-02366a5886a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.265263 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/778cae07-e592-49fd-ac10-02366a5886a6-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.271249 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns" (OuterVolumeSpecName: "kube-api-access-lx9ns") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "kube-api-access-lx9ns". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.310111 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.310288 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config" (OuterVolumeSpecName: "config") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.310833 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.316964 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.319679 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "389c1d0d-ddc1-4113-9bdb-2142af81e18f" (UID: "389c1d0d-ddc1-4113-9bdb-2142af81e18f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355238 4708 generic.go:334] "Generic (PLEG): container finished" podID="778cae07-e592-49fd-ac10-02366a5886a6" containerID="5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" exitCode=0 Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355276 4708 generic.go:334] "Generic (PLEG): container finished" podID="778cae07-e592-49fd-ac10-02366a5886a6" containerID="ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" exitCode=143 Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355333 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerDied","Data":"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b"} Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355374 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerDied","Data":"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f"} Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355387 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"778cae07-e592-49fd-ac10-02366a5886a6","Type":"ContainerDied","Data":"58f2c299726dbbbfa7d4579039c9782ca7bde75f3ce1fafae2bc1440421246e4"} Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355409 4708 scope.go:117] "RemoveContainer" containerID="5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.355742 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.366626 4708 generic.go:334] "Generic (PLEG): container finished" podID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerID="d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56" exitCode=0 Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367128 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" event={"ID":"389c1d0d-ddc1-4113-9bdb-2142af81e18f","Type":"ContainerDied","Data":"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56"} Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367254 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" event={"ID":"389c1d0d-ddc1-4113-9bdb-2142af81e18f","Type":"ContainerDied","Data":"75c78060b6f5d959ef9847f38161736229b9e70c947641668131d5c8a9d5855d"} Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367652 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bf7f4798c-p7n82" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367739 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367899 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367911 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367922 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367932 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lx9ns\" (UniqueName: \"kubernetes.io/projected/389c1d0d-ddc1-4113-9bdb-2142af81e18f-kube-api-access-lx9ns\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.367942 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/389c1d0d-ddc1-4113-9bdb-2142af81e18f-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.370998 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-log" containerID="cri-o://2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" gracePeriod=30 Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.371179 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-metadata" containerID="cri-o://707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" gracePeriod=30 Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.432671 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.445163 4708 scope.go:117] "RemoveContainer" containerID="ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.449964 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.473503 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.479839 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.480302 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="527c5c0e-4e68-4533-b9b0-e65cfd465ce7" containerName="nova-manage" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480323 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="527c5c0e-4e68-4533-b9b0-e65cfd465ce7" containerName="nova-manage" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.480335 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-log" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480342 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-log" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.480349 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-api" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480356 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-api" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.480363 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="dnsmasq-dns" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480368 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="dnsmasq-dns" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.480379 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="init" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480384 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="init" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480579 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="527c5c0e-4e68-4533-b9b0-e65cfd465ce7" containerName="nova-manage" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480609 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-log" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480619 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="778cae07-e592-49fd-ac10-02366a5886a6" containerName="nova-api-api" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.480631 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" containerName="dnsmasq-dns" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.481635 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.484755 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.486751 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bf7f4798c-p7n82"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.488774 4708 scope.go:117] "RemoveContainer" containerID="5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.490360 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b\": container with ID starting with 5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b not found: ID does not exist" containerID="5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.490403 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b"} err="failed to get container status \"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b\": rpc error: code = NotFound desc = could not find container \"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b\": container with ID starting with 5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.490436 4708 scope.go:117] "RemoveContainer" containerID="ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.491784 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f\": container with ID starting with ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f not found: ID does not exist" containerID="ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.491817 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f"} err="failed to get container status \"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f\": rpc error: code = NotFound desc = could not find container \"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f\": container with ID starting with ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.491838 4708 scope.go:117] "RemoveContainer" containerID="5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.492512 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.493060 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b"} err="failed to get container status \"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b\": rpc error: code = NotFound desc = could not find container \"5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b\": container with ID starting with 5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.493104 4708 scope.go:117] "RemoveContainer" containerID="ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.493997 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f"} err="failed to get container status \"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f\": rpc error: code = NotFound desc = could not find container \"ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f\": container with ID starting with ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.494022 4708 scope.go:117] "RemoveContainer" containerID="d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.574842 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.574902 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.575005 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.575061 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2g7zr\" (UniqueName: \"kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.653248 4708 scope.go:117] "RemoveContainer" containerID="572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.677188 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.677304 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.677447 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.677587 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2g7zr\" (UniqueName: \"kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.677776 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.682121 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.682205 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.689640 4708 scope.go:117] "RemoveContainer" containerID="d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.699708 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56\": container with ID starting with d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56 not found: ID does not exist" containerID="d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.699760 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56"} err="failed to get container status \"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56\": rpc error: code = NotFound desc = could not find container \"d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56\": container with ID starting with d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56 not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.699789 4708 scope.go:117] "RemoveContainer" containerID="572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.699982 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2g7zr\" (UniqueName: \"kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr\") pod \"nova-api-0\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: E1125 05:57:04.700662 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16\": container with ID starting with 572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16 not found: ID does not exist" containerID="572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.700708 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16"} err="failed to get container status \"572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16\": rpc error: code = NotFound desc = could not find container \"572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16\": container with ID starting with 572bde7ba75f6cdb29d8435019f390a36f0373f2a1830eeb22a374712e031b16 not found: ID does not exist" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.853221 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.901751 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="389c1d0d-ddc1-4113-9bdb-2142af81e18f" path="/var/lib/kubelet/pods/389c1d0d-ddc1-4113-9bdb-2142af81e18f/volumes" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.902349 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="778cae07-e592-49fd-ac10-02366a5886a6" path="/var/lib/kubelet/pods/778cae07-e592-49fd-ac10-02366a5886a6/volumes" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.936770 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.984329 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w7jr\" (UniqueName: \"kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr\") pod \"473d0c9c-1909-4b9b-932e-17697b75473a\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.984377 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs\") pod \"473d0c9c-1909-4b9b-932e-17697b75473a\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.984581 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle\") pod \"473d0c9c-1909-4b9b-932e-17697b75473a\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.984613 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data\") pod \"473d0c9c-1909-4b9b-932e-17697b75473a\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.984805 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs\") pod \"473d0c9c-1909-4b9b-932e-17697b75473a\" (UID: \"473d0c9c-1909-4b9b-932e-17697b75473a\") " Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.985120 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs" (OuterVolumeSpecName: "logs") pod "473d0c9c-1909-4b9b-932e-17697b75473a" (UID: "473d0c9c-1909-4b9b-932e-17697b75473a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.985637 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/473d0c9c-1909-4b9b-932e-17697b75473a-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:04 crc kubenswrapper[4708]: I1125 05:57:04.994438 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr" (OuterVolumeSpecName: "kube-api-access-5w7jr") pod "473d0c9c-1909-4b9b-932e-17697b75473a" (UID: "473d0c9c-1909-4b9b-932e-17697b75473a"). InnerVolumeSpecName "kube-api-access-5w7jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.009481 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data" (OuterVolumeSpecName: "config-data") pod "473d0c9c-1909-4b9b-932e-17697b75473a" (UID: "473d0c9c-1909-4b9b-932e-17697b75473a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.013114 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "473d0c9c-1909-4b9b-932e-17697b75473a" (UID: "473d0c9c-1909-4b9b-932e-17697b75473a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.048718 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "473d0c9c-1909-4b9b-932e-17697b75473a" (UID: "473d0c9c-1909-4b9b-932e-17697b75473a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.088074 4708 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.088100 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w7jr\" (UniqueName: \"kubernetes.io/projected/473d0c9c-1909-4b9b-932e-17697b75473a-kube-api-access-5w7jr\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.088113 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.088122 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473d0c9c-1909-4b9b-932e-17697b75473a-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379664 4708 generic.go:334] "Generic (PLEG): container finished" podID="473d0c9c-1909-4b9b-932e-17697b75473a" containerID="707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" exitCode=0 Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379709 4708 generic.go:334] "Generic (PLEG): container finished" podID="473d0c9c-1909-4b9b-932e-17697b75473a" containerID="2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" exitCode=143 Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379800 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerDied","Data":"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5"} Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379835 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerDied","Data":"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc"} Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379845 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"473d0c9c-1909-4b9b-932e-17697b75473a","Type":"ContainerDied","Data":"90bd4268717525a33e49757207c8b1745b42436c52320aba36521e1446a24b79"} Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379828 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.379864 4708 scope.go:117] "RemoveContainer" containerID="707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.384587 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.390124 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerName="nova-scheduler-scheduler" containerID="cri-o://10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" gracePeriod=30 Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.435464 4708 scope.go:117] "RemoveContainer" containerID="2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.437210 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.445488 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.451912 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:05 crc kubenswrapper[4708]: E1125 05:57:05.452316 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-log" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.452334 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-log" Nov 25 05:57:05 crc kubenswrapper[4708]: E1125 05:57:05.452352 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-metadata" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.452359 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-metadata" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.453394 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-metadata" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.453423 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" containerName="nova-metadata-log" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.454319 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.460281 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.460931 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.470939 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.512484 4708 scope.go:117] "RemoveContainer" containerID="707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" Nov 25 05:57:05 crc kubenswrapper[4708]: E1125 05:57:05.513027 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5\": container with ID starting with 707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5 not found: ID does not exist" containerID="707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.513064 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5"} err="failed to get container status \"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5\": rpc error: code = NotFound desc = could not find container \"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5\": container with ID starting with 707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5 not found: ID does not exist" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.513090 4708 scope.go:117] "RemoveContainer" containerID="2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" Nov 25 05:57:05 crc kubenswrapper[4708]: E1125 05:57:05.514879 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc\": container with ID starting with 2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc not found: ID does not exist" containerID="2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.514902 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc"} err="failed to get container status \"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc\": rpc error: code = NotFound desc = could not find container \"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc\": container with ID starting with 2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc not found: ID does not exist" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.514917 4708 scope.go:117] "RemoveContainer" containerID="707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.515182 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5"} err="failed to get container status \"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5\": rpc error: code = NotFound desc = could not find container \"707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5\": container with ID starting with 707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5 not found: ID does not exist" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.515206 4708 scope.go:117] "RemoveContainer" containerID="2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.515717 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc"} err="failed to get container status \"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc\": rpc error: code = NotFound desc = could not find container \"2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc\": container with ID starting with 2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc not found: ID does not exist" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.600612 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfwmj\" (UniqueName: \"kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.600824 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.600883 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.600909 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.600990 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.703880 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.704730 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfwmj\" (UniqueName: \"kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.704903 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.705290 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.705384 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.705908 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.707275 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.708427 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.709084 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.719707 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfwmj\" (UniqueName: \"kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj\") pod \"nova-metadata-0\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " pod="openstack/nova-metadata-0" Nov 25 05:57:05 crc kubenswrapper[4708]: I1125 05:57:05.837428 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.278699 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.406304 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerStarted","Data":"5707dab845ce0c607b75cb12b2c2b1e546165f7313e79735846359eb3591857d"} Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.416316 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerStarted","Data":"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71"} Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.416405 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerStarted","Data":"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee"} Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.416499 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerStarted","Data":"1682eb066c32daaafbd60b466a52ca46ad5abb3096edb4e439d55a87bcddd807"} Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.443193 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.4431751 podStartE2EDuration="2.4431751s" podCreationTimestamp="2025-11-25 05:57:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:06.431651442 +0000 UTC m=+967.840484828" watchObservedRunningTime="2025-11-25 05:57:06.4431751 +0000 UTC m=+967.852008476" Nov 25 05:57:06 crc kubenswrapper[4708]: I1125 05:57:06.904965 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="473d0c9c-1909-4b9b-932e-17697b75473a" path="/var/lib/kubelet/pods/473d0c9c-1909-4b9b-932e-17697b75473a/volumes" Nov 25 05:57:07 crc kubenswrapper[4708]: I1125 05:57:07.429787 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerStarted","Data":"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523"} Nov 25 05:57:07 crc kubenswrapper[4708]: I1125 05:57:07.429855 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerStarted","Data":"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db"} Nov 25 05:57:07 crc kubenswrapper[4708]: I1125 05:57:07.453711 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.453695961 podStartE2EDuration="2.453695961s" podCreationTimestamp="2025-11-25 05:57:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:07.446184375 +0000 UTC m=+968.855017761" watchObservedRunningTime="2025-11-25 05:57:07.453695961 +0000 UTC m=+968.862529348" Nov 25 05:57:08 crc kubenswrapper[4708]: E1125 05:57:08.269401 4708 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 05:57:08 crc kubenswrapper[4708]: E1125 05:57:08.271584 4708 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 05:57:08 crc kubenswrapper[4708]: E1125 05:57:08.273248 4708 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 05:57:08 crc kubenswrapper[4708]: E1125 05:57:08.273293 4708 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerName="nova-scheduler-scheduler" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.330049 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.447170 4708 generic.go:334] "Generic (PLEG): container finished" podID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" exitCode=0 Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.447215 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f16b36f-cb7d-4820-808c-e3fccd217e90","Type":"ContainerDied","Data":"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9"} Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.447233 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.447262 4708 scope.go:117] "RemoveContainer" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.447249 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2f16b36f-cb7d-4820-808c-e3fccd217e90","Type":"ContainerDied","Data":"152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0"} Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.468932 4708 scope.go:117] "RemoveContainer" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" Nov 25 05:57:09 crc kubenswrapper[4708]: E1125 05:57:09.469306 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9\": container with ID starting with 10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9 not found: ID does not exist" containerID="10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.469341 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9"} err="failed to get container status \"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9\": rpc error: code = NotFound desc = could not find container \"10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9\": container with ID starting with 10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9 not found: ID does not exist" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.496899 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data\") pod \"2f16b36f-cb7d-4820-808c-e3fccd217e90\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.497197 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tqlh8\" (UniqueName: \"kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8\") pod \"2f16b36f-cb7d-4820-808c-e3fccd217e90\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.497265 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") pod \"2f16b36f-cb7d-4820-808c-e3fccd217e90\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.503828 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8" (OuterVolumeSpecName: "kube-api-access-tqlh8") pod "2f16b36f-cb7d-4820-808c-e3fccd217e90" (UID: "2f16b36f-cb7d-4820-808c-e3fccd217e90"). InnerVolumeSpecName "kube-api-access-tqlh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:09 crc kubenswrapper[4708]: E1125 05:57:09.520500 4708 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle podName:2f16b36f-cb7d-4820-808c-e3fccd217e90 nodeName:}" failed. No retries permitted until 2025-11-25 05:57:10.020447412 +0000 UTC m=+971.429280798 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle") pod "2f16b36f-cb7d-4820-808c-e3fccd217e90" (UID: "2f16b36f-cb7d-4820-808c-e3fccd217e90") : error deleting /var/lib/kubelet/pods/2f16b36f-cb7d-4820-808c-e3fccd217e90/volume-subpaths: remove /var/lib/kubelet/pods/2f16b36f-cb7d-4820-808c-e3fccd217e90/volume-subpaths: no such file or directory Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.522419 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data" (OuterVolumeSpecName: "config-data") pod "2f16b36f-cb7d-4820-808c-e3fccd217e90" (UID: "2f16b36f-cb7d-4820-808c-e3fccd217e90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.601707 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:09 crc kubenswrapper[4708]: I1125 05:57:09.601746 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tqlh8\" (UniqueName: \"kubernetes.io/projected/2f16b36f-cb7d-4820-808c-e3fccd217e90-kube-api-access-tqlh8\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.112621 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") pod \"2f16b36f-cb7d-4820-808c-e3fccd217e90\" (UID: \"2f16b36f-cb7d-4820-808c-e3fccd217e90\") " Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.116787 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f16b36f-cb7d-4820-808c-e3fccd217e90" (UID: "2f16b36f-cb7d-4820-808c-e3fccd217e90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.216090 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f16b36f-cb7d-4820-808c-e3fccd217e90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.381233 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.389624 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.396131 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:10 crc kubenswrapper[4708]: E1125 05:57:10.396634 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerName="nova-scheduler-scheduler" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.396655 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerName="nova-scheduler-scheduler" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.396907 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" containerName="nova-scheduler-scheduler" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.397776 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.399278 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.401022 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.521636 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.522187 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfm4z\" (UniqueName: \"kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.522248 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.625190 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfm4z\" (UniqueName: \"kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.625256 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.625307 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.632229 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.632640 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.639836 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfm4z\" (UniqueName: \"kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z\") pod \"nova-scheduler-0\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.724820 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.837926 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.838189 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:57:10 crc kubenswrapper[4708]: I1125 05:57:10.924064 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f16b36f-cb7d-4820-808c-e3fccd217e90" path="/var/lib/kubelet/pods/2f16b36f-cb7d-4820-808c-e3fccd217e90/volumes" Nov 25 05:57:11 crc kubenswrapper[4708]: I1125 05:57:11.121957 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:11 crc kubenswrapper[4708]: I1125 05:57:11.474532 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50","Type":"ContainerStarted","Data":"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d"} Nov 25 05:57:11 crc kubenswrapper[4708]: I1125 05:57:11.474865 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50","Type":"ContainerStarted","Data":"b79fe53fbfab20a3623ea556617ffbf1e3e53785b30fed3d38fcb9687bf8cbf7"} Nov 25 05:57:11 crc kubenswrapper[4708]: I1125 05:57:11.499491 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.499476606 podStartE2EDuration="1.499476606s" podCreationTimestamp="2025-11-25 05:57:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:11.497825093 +0000 UTC m=+972.906658479" watchObservedRunningTime="2025-11-25 05:57:11.499476606 +0000 UTC m=+972.908309992" Nov 25 05:57:12 crc kubenswrapper[4708]: I1125 05:57:12.735059 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 05:57:14 crc kubenswrapper[4708]: I1125 05:57:14.937367 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:57:14 crc kubenswrapper[4708]: I1125 05:57:14.937786 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:57:15 crc kubenswrapper[4708]: I1125 05:57:15.725392 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 05:57:15 crc kubenswrapper[4708]: I1125 05:57:15.837905 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 05:57:15 crc kubenswrapper[4708]: I1125 05:57:15.837967 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 05:57:16 crc kubenswrapper[4708]: I1125 05:57:16.019725 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:16 crc kubenswrapper[4708]: I1125 05:57:16.019758 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:16 crc kubenswrapper[4708]: I1125 05:57:16.852649 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:16 crc kubenswrapper[4708]: I1125 05:57:16.852692 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:20 crc kubenswrapper[4708]: I1125 05:57:20.725890 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 05:57:20 crc kubenswrapper[4708]: I1125 05:57:20.752851 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 05:57:21 crc kubenswrapper[4708]: I1125 05:57:21.606167 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 05:57:24 crc kubenswrapper[4708]: I1125 05:57:24.943265 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 05:57:24 crc kubenswrapper[4708]: I1125 05:57:24.943782 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 05:57:24 crc kubenswrapper[4708]: I1125 05:57:24.944381 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 05:57:24 crc kubenswrapper[4708]: I1125 05:57:24.947818 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.625673 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.631691 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.792444 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.797456 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.806866 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.843704 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.846629 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.853720 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857197 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dntx8\" (UniqueName: \"kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857264 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857388 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857502 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857701 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.857756 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.959902 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.959955 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.960028 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dntx8\" (UniqueName: \"kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.960096 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.960135 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.960177 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.961381 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.961397 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.961636 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.961702 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.961835 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:25 crc kubenswrapper[4708]: I1125 05:57:25.977955 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dntx8\" (UniqueName: \"kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8\") pod \"dnsmasq-dns-bc466d6bf-tqhcq\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:26 crc kubenswrapper[4708]: I1125 05:57:26.118487 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:26 crc kubenswrapper[4708]: I1125 05:57:26.597318 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:57:26 crc kubenswrapper[4708]: I1125 05:57:26.599777 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 05:57:26 crc kubenswrapper[4708]: I1125 05:57:26.640432 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" event={"ID":"9aa2e102-5381-478f-91ef-8e0fe9ec13fb","Type":"ContainerStarted","Data":"4167e595767e030e8ac4d79111a4aa1ac492ec2ed82758a9831af241bba5b2f7"} Nov 25 05:57:26 crc kubenswrapper[4708]: I1125 05:57:26.649510 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 05:57:27 crc kubenswrapper[4708]: I1125 05:57:27.649855 4708 generic.go:334] "Generic (PLEG): container finished" podID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerID="bc97568de5f8bab32b4f9f22974bd97578b8ce69e2dd73432e7a4ccedd824a9c" exitCode=0 Nov 25 05:57:27 crc kubenswrapper[4708]: I1125 05:57:27.650007 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" event={"ID":"9aa2e102-5381-478f-91ef-8e0fe9ec13fb","Type":"ContainerDied","Data":"bc97568de5f8bab32b4f9f22974bd97578b8ce69e2dd73432e7a4ccedd824a9c"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.227909 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.229794 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-central-agent" containerID="cri-o://0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.229875 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-notification-agent" containerID="cri-o://1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.229991 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="sg-core" containerID="cri-o://1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.230061 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="proxy-httpd" containerID="cri-o://41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: W1125 05:57:28.293274 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice/crio-2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc.scope WatchSource:0}: Error finding container 2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc: Status 404 returned error can't find the container with id 2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc Nov 25 05:57:28 crc kubenswrapper[4708]: W1125 05:57:28.294305 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice/crio-707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5.scope WatchSource:0}: Error finding container 707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5: Status 404 returned error can't find the container with id 707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.359644 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:28 crc kubenswrapper[4708]: E1125 05:57:28.572489 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f16b36f_cb7d_4820_808c_e3fccd217e90.slice/crio-152b962804a3bfb5b1d8cb2090af8e7bc5a5876c2f4043601e79d3c1642e78f0\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice/crio-conmon-707ecc90d9f73c5389eb2075070f80f9f27c5160ca452e260a13cfe8fa66acf5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f16b36f_cb7d_4820_808c_e3fccd217e90.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice/crio-58f2c299726dbbbfa7d4579039c9782ca7bde75f3ce1fafae2bc1440421246e4\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice/crio-conmon-ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod389c1d0d_ddc1_4113_9bdb_2142af81e18f.slice/crio-conmon-d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod389c1d0d_ddc1_4113_9bdb_2142af81e18f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d8a5ee_d7b5_462e_88cd_9773753828c1.slice/crio-conmon-41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d8a5ee_d7b5_462e_88cd_9773753828c1.slice/crio-conmon-1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice/crio-ec1efb9412258bf1ccbea7d4e811a3ecb716682f6f931ec9d38597fd4379116f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice/crio-5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod778cae07_e592_49fd_ac10_02366a5886a6.slice/crio-conmon-5c1508df8c414e0ff636f0667cc07db32178961788d94607f56399d55826646b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6c52ba9_bea9_4756_9c31_f2a374c71674.slice/crio-conmon-f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f16b36f_cb7d_4820_808c_e3fccd217e90.slice/crio-10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod527c5c0e_4e68_4533_b9b0_e65cfd465ce7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod389c1d0d_ddc1_4113_9bdb_2142af81e18f.slice/crio-d5f9708bb4eccef8df95724afdd23e512f274a9e6477d761c7cd1c30d2514e56.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d8a5ee_d7b5_462e_88cd_9773753828c1.slice/crio-41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f16b36f_cb7d_4820_808c_e3fccd217e90.slice/crio-conmon-10d69d277b1996c6f51fbaf2b02b13b25a8d05ef337ae679d38e972910d3ffa9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6c52ba9_bea9_4756_9c31_f2a374c71674.slice/crio-f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1d8a5ee_d7b5_462e_88cd_9773753828c1.slice/crio-1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee4549fd_ebcf_4169_b8cf_04ab1e995f9f.slice/crio-34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice/crio-90bd4268717525a33e49757207c8b1745b42436c52320aba36521e1446a24b79\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod473d0c9c_1909_4b9b_932e_17697b75473a.slice/crio-conmon-2b872c4633e060dce40fee9f8b8f32f4f0a42fe55762ad33d16639172b0187dc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod389c1d0d_ddc1_4113_9bdb_2142af81e18f.slice/crio-75c78060b6f5d959ef9847f38161736229b9e70c947641668131d5c8a9d5855d\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod527c5c0e_4e68_4533_b9b0_e65cfd465ce7.slice/crio-ab011b7dd4147eef07d453607a8c584a672d5e4d861c5add16a96d9811bb145b\": RecentStats: unable to find data in memory cache]" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.612605 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.638923 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data\") pod \"d6c52ba9-bea9-4756-9c31-f2a374c71674\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.638971 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle\") pod \"d6c52ba9-bea9-4756-9c31-f2a374c71674\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.639293 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxfxl\" (UniqueName: \"kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl\") pod \"d6c52ba9-bea9-4756-9c31-f2a374c71674\" (UID: \"d6c52ba9-bea9-4756-9c31-f2a374c71674\") " Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.646938 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl" (OuterVolumeSpecName: "kube-api-access-rxfxl") pod "d6c52ba9-bea9-4756-9c31-f2a374c71674" (UID: "d6c52ba9-bea9-4756-9c31-f2a374c71674"). InnerVolumeSpecName "kube-api-access-rxfxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.663996 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data" (OuterVolumeSpecName: "config-data") pod "d6c52ba9-bea9-4756-9c31-f2a374c71674" (UID: "d6c52ba9-bea9-4756-9c31-f2a374c71674"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.664142 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" event={"ID":"9aa2e102-5381-478f-91ef-8e0fe9ec13fb","Type":"ContainerStarted","Data":"6c24c11eeb7b7d84c99b5fd3969b07626761c562aa0c327a656b14e0a49ab46b"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.664299 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.665997 4708 generic.go:334] "Generic (PLEG): container finished" podID="d6c52ba9-bea9-4756-9c31-f2a374c71674" containerID="f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181" exitCode=137 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.666071 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6c52ba9-bea9-4756-9c31-f2a374c71674","Type":"ContainerDied","Data":"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.666108 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"d6c52ba9-bea9-4756-9c31-f2a374c71674","Type":"ContainerDied","Data":"58be7982844aa6d26a7b1b7f3a8501ebe0e6ef02c0f47d331306c49f74b331a9"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.666128 4708 scope.go:117] "RemoveContainer" containerID="f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.666271 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674167 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerID="41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e" exitCode=0 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674221 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerID="1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7" exitCode=2 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674231 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerID="0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df" exitCode=0 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674251 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerDied","Data":"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674279 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerDied","Data":"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674294 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerDied","Data":"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df"} Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674644 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-log" containerID="cri-o://9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.674662 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-api" containerID="cri-o://bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71" gracePeriod=30 Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.677011 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d6c52ba9-bea9-4756-9c31-f2a374c71674" (UID: "d6c52ba9-bea9-4756-9c31-f2a374c71674"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.685161 4708 scope.go:117] "RemoveContainer" containerID="f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181" Nov 25 05:57:28 crc kubenswrapper[4708]: E1125 05:57:28.685494 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181\": container with ID starting with f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181 not found: ID does not exist" containerID="f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.685558 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181"} err="failed to get container status \"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181\": rpc error: code = NotFound desc = could not find container \"f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181\": container with ID starting with f9a350e7c0ed714d6228c3d9b0940c4f5b5140fca09ad3f3ef3e58f240580181 not found: ID does not exist" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.696999 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" podStartSLOduration=3.696983967 podStartE2EDuration="3.696983967s" podCreationTimestamp="2025-11-25 05:57:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:28.679134475 +0000 UTC m=+990.087967861" watchObservedRunningTime="2025-11-25 05:57:28.696983967 +0000 UTC m=+990.105817353" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.741113 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxfxl\" (UniqueName: \"kubernetes.io/projected/d6c52ba9-bea9-4756-9c31-f2a374c71674-kube-api-access-rxfxl\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.741142 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.741152 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6c52ba9-bea9-4756-9c31-f2a374c71674-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:28 crc kubenswrapper[4708]: I1125 05:57:28.993321 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.005917 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.011368 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:57:29 crc kubenswrapper[4708]: E1125 05:57:29.011844 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6c52ba9-bea9-4756-9c31-f2a374c71674" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.011864 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6c52ba9-bea9-4756-9c31-f2a374c71674" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.012096 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6c52ba9-bea9-4756-9c31-f2a374c71674" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.012838 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.014796 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.017125 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.019262 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.031838 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.046182 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.046234 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.046276 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.046340 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck494\" (UniqueName: \"kubernetes.io/projected/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-kube-api-access-ck494\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.046393 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.148842 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.148905 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.148954 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.149034 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck494\" (UniqueName: \"kubernetes.io/projected/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-kube-api-access-ck494\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.149094 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.153179 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.153270 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.153975 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.155873 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.165435 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck494\" (UniqueName: \"kubernetes.io/projected/e9ad8e31-f124-4cc8-ab77-ff50df6589f1-kube-api-access-ck494\") pod \"nova-cell1-novncproxy-0\" (UID: \"e9ad8e31-f124-4cc8-ab77-ff50df6589f1\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.332046 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.687689 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerID="9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee" exitCode=143 Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.687893 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerDied","Data":"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee"} Nov 25 05:57:29 crc kubenswrapper[4708]: I1125 05:57:29.774663 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 05:57:29 crc kubenswrapper[4708]: W1125 05:57:29.776176 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9ad8e31_f124_4cc8_ab77_ff50df6589f1.slice/crio-56b6d5d33d5a2b2e327f2a1244b7220b3e49b1d83a6b8f7b9e35ac4b8f5e57c8 WatchSource:0}: Error finding container 56b6d5d33d5a2b2e327f2a1244b7220b3e49b1d83a6b8f7b9e35ac4b8f5e57c8: Status 404 returned error can't find the container with id 56b6d5d33d5a2b2e327f2a1244b7220b3e49b1d83a6b8f7b9e35ac4b8f5e57c8 Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.645854 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.646335 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" containerName="kube-state-metrics" containerID="cri-o://75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452" gracePeriod=30 Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.704631 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9ad8e31-f124-4cc8-ab77-ff50df6589f1","Type":"ContainerStarted","Data":"6223716f714e3aeae9c52e27337d232391185edb0c152b3846d302b759d21984"} Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.704696 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e9ad8e31-f124-4cc8-ab77-ff50df6589f1","Type":"ContainerStarted","Data":"56b6d5d33d5a2b2e327f2a1244b7220b3e49b1d83a6b8f7b9e35ac4b8f5e57c8"} Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.730321 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.730302162 podStartE2EDuration="2.730302162s" podCreationTimestamp="2025-11-25 05:57:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:30.72081488 +0000 UTC m=+992.129648257" watchObservedRunningTime="2025-11-25 05:57:30.730302162 +0000 UTC m=+992.139135548" Nov 25 05:57:30 crc kubenswrapper[4708]: I1125 05:57:30.923601 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6c52ba9-bea9-4756-9c31-f2a374c71674" path="/var/lib/kubelet/pods/d6c52ba9-bea9-4756-9c31-f2a374c71674/volumes" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.191388 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.201584 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dw7x\" (UniqueName: \"kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x\") pod \"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460\" (UID: \"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.210782 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x" (OuterVolumeSpecName: "kube-api-access-6dw7x") pod "0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" (UID: "0dbf1bb3-7a60-4c3e-a942-cab2c9a57460"). InnerVolumeSpecName "kube-api-access-6dw7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.306997 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dw7x\" (UniqueName: \"kubernetes.io/projected/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460-kube-api-access-6dw7x\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.474436 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.614929 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.615003 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkz72\" (UniqueName: \"kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.615392 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.615463 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.615987 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.616101 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.616554 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.616575 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd\") pod \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\" (UID: \"e1d8a5ee-d7b5-462e-88cd-9773753828c1\") " Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.616890 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.617146 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.617172 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e1d8a5ee-d7b5-462e-88cd-9773753828c1-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.619297 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72" (OuterVolumeSpecName: "kube-api-access-qkz72") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "kube-api-access-qkz72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.629661 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts" (OuterVolumeSpecName: "scripts") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.641419 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.683494 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.701015 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data" (OuterVolumeSpecName: "config-data") pod "e1d8a5ee-d7b5-462e-88cd-9773753828c1" (UID: "e1d8a5ee-d7b5-462e-88cd-9773753828c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.713841 4708 generic.go:334] "Generic (PLEG): container finished" podID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" containerID="75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452" exitCode=2 Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.713911 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.713917 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460","Type":"ContainerDied","Data":"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452"} Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.714016 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"0dbf1bb3-7a60-4c3e-a942-cab2c9a57460","Type":"ContainerDied","Data":"ced89f359db24b22d05f83622be3d9d37d9d23bff403a31a69e0b72efbcb67f8"} Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.714066 4708 scope.go:117] "RemoveContainer" containerID="75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.718002 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerID="1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e" exitCode=0 Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.718130 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerDied","Data":"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e"} Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.718195 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e1d8a5ee-d7b5-462e-88cd-9773753828c1","Type":"ContainerDied","Data":"86120dae20bc46177936695c106a0a42139200feacfc2756fa01d97c05b8e617"} Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.718271 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.718816 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.719352 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkz72\" (UniqueName: \"kubernetes.io/projected/e1d8a5ee-d7b5-462e-88cd-9773753828c1-kube-api-access-qkz72\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.719399 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.719411 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.719422 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1d8a5ee-d7b5-462e-88cd-9773753828c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.733610 4708 scope.go:117] "RemoveContainer" containerID="75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.735982 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452\": container with ID starting with 75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452 not found: ID does not exist" containerID="75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.736048 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452"} err="failed to get container status \"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452\": rpc error: code = NotFound desc = could not find container \"75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452\": container with ID starting with 75161c5f098f4938dd24300e0cda03cd104d27b591b855e1d4fb449c8209d452 not found: ID does not exist" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.736074 4708 scope.go:117] "RemoveContainer" containerID="41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.743954 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.751587 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.763054 4708 scope.go:117] "RemoveContainer" containerID="1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.768883 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.769296 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-central-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769315 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-central-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.769332 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="proxy-httpd" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769339 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="proxy-httpd" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.769367 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="sg-core" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769373 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="sg-core" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.769384 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-notification-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769390 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-notification-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.769403 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" containerName="kube-state-metrics" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769407 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" containerName="kube-state-metrics" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769629 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="sg-core" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769646 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" containerName="kube-state-metrics" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769659 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="proxy-httpd" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769672 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-notification-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.769682 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" containerName="ceilometer-central-agent" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.770320 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.774073 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.774262 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-k6bw7" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.774390 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.793658 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.793880 4708 scope.go:117] "RemoveContainer" containerID="1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.805087 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.812779 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.815279 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.815677 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.819015 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.819073 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821482 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821575 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821615 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821694 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821760 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821827 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821886 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821912 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821944 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.821995 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqtgx\" (UniqueName: \"kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.822047 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j24j\" (UniqueName: \"kubernetes.io/projected/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-api-access-6j24j\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.822076 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.827013 4708 scope.go:117] "RemoveContainer" containerID="0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.848142 4708 scope.go:117] "RemoveContainer" containerID="41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.848416 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e\": container with ID starting with 41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e not found: ID does not exist" containerID="41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.848460 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e"} err="failed to get container status \"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e\": rpc error: code = NotFound desc = could not find container \"41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e\": container with ID starting with 41ac1f86c8b5caace1e734177d764554078819dc5b74ad768ac622da46bdd39e not found: ID does not exist" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.848493 4708 scope.go:117] "RemoveContainer" containerID="1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.848795 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7\": container with ID starting with 1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7 not found: ID does not exist" containerID="1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.848826 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7"} err="failed to get container status \"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7\": rpc error: code = NotFound desc = could not find container \"1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7\": container with ID starting with 1af2cee44573661a3d0511c4f9b4b3c6141957233c32f88007cc2585b9b0dbc7 not found: ID does not exist" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.848851 4708 scope.go:117] "RemoveContainer" containerID="1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.849057 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e\": container with ID starting with 1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e not found: ID does not exist" containerID="1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.849080 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e"} err="failed to get container status \"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e\": rpc error: code = NotFound desc = could not find container \"1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e\": container with ID starting with 1e94d16bf577d9add2c4853b41ee7635f5fc882c09973dfd6281ad705c0bbc0e not found: ID does not exist" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.849094 4708 scope.go:117] "RemoveContainer" containerID="0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df" Nov 25 05:57:31 crc kubenswrapper[4708]: E1125 05:57:31.849270 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df\": container with ID starting with 0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df not found: ID does not exist" containerID="0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.849293 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df"} err="failed to get container status \"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df\": rpc error: code = NotFound desc = could not find container \"0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df\": container with ID starting with 0d8ab238601de1f155bdabecbe505fdfe75a78d5896e2a5348777ce98b3011df not found: ID does not exist" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.924624 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqtgx\" (UniqueName: \"kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.924692 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j24j\" (UniqueName: \"kubernetes.io/projected/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-api-access-6j24j\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.924930 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925275 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925364 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925602 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925714 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925828 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.925963 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.926005 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.926042 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.926758 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.929636 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.930931 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.931303 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.932400 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.933839 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.938711 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.939787 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j24j\" (UniqueName: \"kubernetes.io/projected/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-api-access-6j24j\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.945926 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.955078 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f6e8f862-6f67-426c-8e31-cb1f6baaa44e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f6e8f862-6f67-426c-8e31-cb1f6baaa44e\") " pod="openstack/kube-state-metrics-0" Nov 25 05:57:31 crc kubenswrapper[4708]: I1125 05:57:31.957051 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqtgx\" (UniqueName: \"kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx\") pod \"ceilometer-0\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " pod="openstack/ceilometer-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.105022 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.137450 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.196730 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.233482 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2g7zr\" (UniqueName: \"kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr\") pod \"c5815599-b3c9-4cf6-9929-6fdb90d45439\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.233969 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle\") pod \"c5815599-b3c9-4cf6-9929-6fdb90d45439\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.233998 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs\") pod \"c5815599-b3c9-4cf6-9929-6fdb90d45439\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.234185 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data\") pod \"c5815599-b3c9-4cf6-9929-6fdb90d45439\" (UID: \"c5815599-b3c9-4cf6-9929-6fdb90d45439\") " Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.235626 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs" (OuterVolumeSpecName: "logs") pod "c5815599-b3c9-4cf6-9929-6fdb90d45439" (UID: "c5815599-b3c9-4cf6-9929-6fdb90d45439"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.240254 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr" (OuterVolumeSpecName: "kube-api-access-2g7zr") pod "c5815599-b3c9-4cf6-9929-6fdb90d45439" (UID: "c5815599-b3c9-4cf6-9929-6fdb90d45439"). InnerVolumeSpecName "kube-api-access-2g7zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.260757 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c5815599-b3c9-4cf6-9929-6fdb90d45439" (UID: "c5815599-b3c9-4cf6-9929-6fdb90d45439"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.266829 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data" (OuterVolumeSpecName: "config-data") pod "c5815599-b3c9-4cf6-9929-6fdb90d45439" (UID: "c5815599-b3c9-4cf6-9929-6fdb90d45439"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.336021 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.336054 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c5815599-b3c9-4cf6-9929-6fdb90d45439-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.336065 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c5815599-b3c9-4cf6-9929-6fdb90d45439-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.336074 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2g7zr\" (UniqueName: \"kubernetes.io/projected/c5815599-b3c9-4cf6-9929-6fdb90d45439-kube-api-access-2g7zr\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.605894 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.718833 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: W1125 05:57:32.728207 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod552327a3_d3e4_4e09_8545_fae3a15a52ca.slice/crio-37c6b2a32ad87f14ddb29837a4a5535789e9b9b59a6806864c4498b40537ad9b WatchSource:0}: Error finding container 37c6b2a32ad87f14ddb29837a4a5535789e9b9b59a6806864c4498b40537ad9b: Status 404 returned error can't find the container with id 37c6b2a32ad87f14ddb29837a4a5535789e9b9b59a6806864c4498b40537ad9b Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.731918 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerID="bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71" exitCode=0 Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.732009 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.732048 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerDied","Data":"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71"} Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.732084 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c5815599-b3c9-4cf6-9929-6fdb90d45439","Type":"ContainerDied","Data":"1682eb066c32daaafbd60b466a52ca46ad5abb3096edb4e439d55a87bcddd807"} Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.732104 4708 scope.go:117] "RemoveContainer" containerID="bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.736429 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f6e8f862-6f67-426c-8e31-cb1f6baaa44e","Type":"ContainerStarted","Data":"3648020280c3699edd515ad94444182cf4fe9bd5749bae9a5b0e4a8135290c18"} Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.780952 4708 scope.go:117] "RemoveContainer" containerID="9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.787297 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.801666 4708 scope.go:117] "RemoveContainer" containerID="bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71" Nov 25 05:57:32 crc kubenswrapper[4708]: E1125 05:57:32.801992 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71\": container with ID starting with bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71 not found: ID does not exist" containerID="bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.802022 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71"} err="failed to get container status \"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71\": rpc error: code = NotFound desc = could not find container \"bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71\": container with ID starting with bd5071e4910f4b7b7d1ac104cbf1a4921405b7ef0489952c48e986f768ad1e71 not found: ID does not exist" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.802039 4708 scope.go:117] "RemoveContainer" containerID="9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee" Nov 25 05:57:32 crc kubenswrapper[4708]: E1125 05:57:32.802332 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee\": container with ID starting with 9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee not found: ID does not exist" containerID="9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.802367 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee"} err="failed to get container status \"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee\": rpc error: code = NotFound desc = could not find container \"9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee\": container with ID starting with 9bb9b11087a4f7dcbddf52bceff6ccdd76078b3218ce4540b12950fd6850dbee not found: ID does not exist" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.806458 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.813165 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: E1125 05:57:32.813715 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-log" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.813736 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-log" Nov 25 05:57:32 crc kubenswrapper[4708]: E1125 05:57:32.813751 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-api" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.813758 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-api" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.813959 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-log" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.813982 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" containerName="nova-api-api" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.815129 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.818678 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.819384 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.819696 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.819880 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.902918 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dbf1bb3-7a60-4c3e-a942-cab2c9a57460" path="/var/lib/kubelet/pods/0dbf1bb3-7a60-4c3e-a942-cab2c9a57460/volumes" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.903457 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5815599-b3c9-4cf6-9929-6fdb90d45439" path="/var/lib/kubelet/pods/c5815599-b3c9-4cf6-9929-6fdb90d45439/volumes" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.903997 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d8a5ee-d7b5-462e-88cd-9773753828c1" path="/var/lib/kubelet/pods/e1d8a5ee-d7b5-462e-88cd-9773753828c1/volumes" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.946544 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szkdp\" (UniqueName: \"kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.946841 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.946889 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.946912 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.946971 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:32 crc kubenswrapper[4708]: I1125 05:57:32.947341 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049485 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szkdp\" (UniqueName: \"kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049638 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049762 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049781 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049875 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.049916 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.052068 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.054692 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.055718 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.055887 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.056394 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.066481 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szkdp\" (UniqueName: \"kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp\") pod \"nova-api-0\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.130196 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.224601 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.564259 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:33 crc kubenswrapper[4708]: W1125 05:57:33.571434 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod03ec42e1_aa5d_446f_ac4a_de7559fde723.slice/crio-2f8cf7a077264e17c615858933545917c319b33cee3832939468b155a6471190 WatchSource:0}: Error finding container 2f8cf7a077264e17c615858933545917c319b33cee3832939468b155a6471190: Status 404 returned error can't find the container with id 2f8cf7a077264e17c615858933545917c319b33cee3832939468b155a6471190 Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.750236 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f6e8f862-6f67-426c-8e31-cb1f6baaa44e","Type":"ContainerStarted","Data":"fd1a33384a2d6590a0df0e8503f10d05d0b5f8cb42e64b6981f122ae7e5411fd"} Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.750613 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.754238 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerStarted","Data":"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516"} Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.754272 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerStarted","Data":"2f8cf7a077264e17c615858933545917c319b33cee3832939468b155a6471190"} Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.756502 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerStarted","Data":"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582"} Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.756574 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerStarted","Data":"37c6b2a32ad87f14ddb29837a4a5535789e9b9b59a6806864c4498b40537ad9b"} Nov 25 05:57:33 crc kubenswrapper[4708]: I1125 05:57:33.776486 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.475316883 podStartE2EDuration="2.776408181s" podCreationTimestamp="2025-11-25 05:57:31 +0000 UTC" firstStartedPulling="2025-11-25 05:57:32.616195504 +0000 UTC m=+994.025028890" lastFinishedPulling="2025-11-25 05:57:32.917286802 +0000 UTC m=+994.326120188" observedRunningTime="2025-11-25 05:57:33.764604395 +0000 UTC m=+995.173437781" watchObservedRunningTime="2025-11-25 05:57:33.776408181 +0000 UTC m=+995.185241567" Nov 25 05:57:34 crc kubenswrapper[4708]: I1125 05:57:34.333646 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:34 crc kubenswrapper[4708]: I1125 05:57:34.768417 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerStarted","Data":"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379"} Nov 25 05:57:34 crc kubenswrapper[4708]: I1125 05:57:34.771982 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerStarted","Data":"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938"} Nov 25 05:57:34 crc kubenswrapper[4708]: I1125 05:57:34.792772 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.792747591 podStartE2EDuration="2.792747591s" podCreationTimestamp="2025-11-25 05:57:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:34.78493699 +0000 UTC m=+996.193770376" watchObservedRunningTime="2025-11-25 05:57:34.792747591 +0000 UTC m=+996.201580977" Nov 25 05:57:35 crc kubenswrapper[4708]: I1125 05:57:35.791741 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerStarted","Data":"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076"} Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.120788 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.177070 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.177345 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-d6558c9-hghrr" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="dnsmasq-dns" containerID="cri-o://62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448" gracePeriod=10 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.617045 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642100 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642258 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tws7\" (UniqueName: \"kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642332 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642375 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642416 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.642455 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config\") pod \"d954cc45-5468-4663-ae00-9076e5ecae8e\" (UID: \"d954cc45-5468-4663-ae00-9076e5ecae8e\") " Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.649796 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7" (OuterVolumeSpecName: "kube-api-access-8tws7") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "kube-api-access-8tws7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.703098 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config" (OuterVolumeSpecName: "config") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.703972 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.708917 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.713882 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.737044 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d954cc45-5468-4663-ae00-9076e5ecae8e" (UID: "d954cc45-5468-4663-ae00-9076e5ecae8e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745411 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745439 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tws7\" (UniqueName: \"kubernetes.io/projected/d954cc45-5468-4663-ae00-9076e5ecae8e-kube-api-access-8tws7\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745450 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745460 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745473 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.745482 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d954cc45-5468-4663-ae00-9076e5ecae8e-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.801592 4708 generic.go:334] "Generic (PLEG): container finished" podID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerID="8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226" exitCode=1 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.801661 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerDied","Data":"8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226"} Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.801714 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-central-agent" containerID="cri-o://98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582" gracePeriod=30 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.801763 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="sg-core" containerID="cri-o://110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076" gracePeriod=30 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.801794 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-notification-agent" containerID="cri-o://22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938" gracePeriod=30 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.804806 4708 generic.go:334] "Generic (PLEG): container finished" podID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerID="62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448" exitCode=0 Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.804847 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6558c9-hghrr" event={"ID":"d954cc45-5468-4663-ae00-9076e5ecae8e","Type":"ContainerDied","Data":"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448"} Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.804877 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-d6558c9-hghrr" event={"ID":"d954cc45-5468-4663-ae00-9076e5ecae8e","Type":"ContainerDied","Data":"15e4ee1fa7f63d764b4c2bb728cebd995b32a4749d0976f9ae9f8846f68e222f"} Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.804886 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-d6558c9-hghrr" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.804894 4708 scope.go:117] "RemoveContainer" containerID="62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.832485 4708 scope.go:117] "RemoveContainer" containerID="e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.845118 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.851207 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-d6558c9-hghrr"] Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.862753 4708 scope.go:117] "RemoveContainer" containerID="62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448" Nov 25 05:57:36 crc kubenswrapper[4708]: E1125 05:57:36.863133 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448\": container with ID starting with 62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448 not found: ID does not exist" containerID="62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.863166 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448"} err="failed to get container status \"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448\": rpc error: code = NotFound desc = could not find container \"62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448\": container with ID starting with 62a12820af3a444bbfadf6e5a17a5881354fc283c25d36b725d52bc5c0ae1448 not found: ID does not exist" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.863190 4708 scope.go:117] "RemoveContainer" containerID="e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592" Nov 25 05:57:36 crc kubenswrapper[4708]: E1125 05:57:36.863597 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592\": container with ID starting with e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592 not found: ID does not exist" containerID="e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.863629 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592"} err="failed to get container status \"e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592\": rpc error: code = NotFound desc = could not find container \"e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592\": container with ID starting with e9a221cc0968f3be418905da60bf42665b53d8fa92fef9890928b46184308592 not found: ID does not exist" Nov 25 05:57:36 crc kubenswrapper[4708]: I1125 05:57:36.914161 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" path="/var/lib/kubelet/pods/d954cc45-5468-4663-ae00-9076e5ecae8e/volumes" Nov 25 05:57:37 crc kubenswrapper[4708]: I1125 05:57:37.815292 4708 generic.go:334] "Generic (PLEG): container finished" podID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerID="110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076" exitCode=2 Nov 25 05:57:37 crc kubenswrapper[4708]: I1125 05:57:37.815327 4708 generic.go:334] "Generic (PLEG): container finished" podID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerID="22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938" exitCode=0 Nov 25 05:57:37 crc kubenswrapper[4708]: I1125 05:57:37.815364 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerDied","Data":"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076"} Nov 25 05:57:37 crc kubenswrapper[4708]: I1125 05:57:37.815399 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerDied","Data":"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938"} Nov 25 05:57:38 crc kubenswrapper[4708]: E1125 05:57:38.814179 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee4549fd_ebcf_4169_b8cf_04ab1e995f9f.slice/crio-34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925\": RecentStats: unable to find data in memory cache]" Nov 25 05:57:39 crc kubenswrapper[4708]: I1125 05:57:39.332973 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:39 crc kubenswrapper[4708]: I1125 05:57:39.352130 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:39 crc kubenswrapper[4708]: I1125 05:57:39.852395 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.063512 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-d96zq"] Nov 25 05:57:40 crc kubenswrapper[4708]: E1125 05:57:40.064179 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="init" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.064200 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="init" Nov 25 05:57:40 crc kubenswrapper[4708]: E1125 05:57:40.064214 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="dnsmasq-dns" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.064220 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="dnsmasq-dns" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.064420 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d954cc45-5468-4663-ae00-9076e5ecae8e" containerName="dnsmasq-dns" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.065245 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.067968 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.068201 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.075099 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d96zq"] Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.215296 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.215393 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.215761 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hgcz\" (UniqueName: \"kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.215915 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.318384 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hgcz\" (UniqueName: \"kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.318491 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.318591 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.318625 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.335095 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.335281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.339199 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.346041 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hgcz\" (UniqueName: \"kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz\") pod \"nova-cell1-cell-mapping-d96zq\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.384311 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.830018 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-d96zq"] Nov 25 05:57:40 crc kubenswrapper[4708]: W1125 05:57:40.832047 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcee0a911_c6e0_43b0_9b11_81f4d283dea4.slice/crio-8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1 WatchSource:0}: Error finding container 8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1: Status 404 returned error can't find the container with id 8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1 Nov 25 05:57:40 crc kubenswrapper[4708]: I1125 05:57:40.848641 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d96zq" event={"ID":"cee0a911-c6e0-43b0-9b11-81f4d283dea4","Type":"ContainerStarted","Data":"8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1"} Nov 25 05:57:41 crc kubenswrapper[4708]: I1125 05:57:41.859746 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d96zq" event={"ID":"cee0a911-c6e0-43b0-9b11-81f4d283dea4","Type":"ContainerStarted","Data":"961cb4a067c64ddf40f3839f15f99e2ef29c9d639cb758cc8daf5b4deb9bef7b"} Nov 25 05:57:41 crc kubenswrapper[4708]: I1125 05:57:41.878326 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-d96zq" podStartSLOduration=1.8783096910000001 podStartE2EDuration="1.878309691s" podCreationTimestamp="2025-11-25 05:57:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:41.877600836 +0000 UTC m=+1003.286434222" watchObservedRunningTime="2025-11-25 05:57:41.878309691 +0000 UTC m=+1003.287143077" Nov 25 05:57:42 crc kubenswrapper[4708]: I1125 05:57:42.118669 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.131916 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.132405 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.310774 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502233 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqtgx\" (UniqueName: \"kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502306 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502532 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502572 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502666 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502759 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502863 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data\") pod \"552327a3-d3e4-4e09-8545-fae3a15a52ca\" (UID: \"552327a3-d3e4-4e09-8545-fae3a15a52ca\") " Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.502991 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.503217 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.503789 4708 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.503812 4708 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/552327a3-d3e4-4e09-8545-fae3a15a52ca-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.510743 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx" (OuterVolumeSpecName: "kube-api-access-dqtgx") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "kube-api-access-dqtgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.526264 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts" (OuterVolumeSpecName: "scripts") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.543640 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.571651 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.588941 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data" (OuterVolumeSpecName: "config-data") pod "552327a3-d3e4-4e09-8545-fae3a15a52ca" (UID: "552327a3-d3e4-4e09-8545-fae3a15a52ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.604941 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.605028 4708 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.605101 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.605157 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqtgx\" (UniqueName: \"kubernetes.io/projected/552327a3-d3e4-4e09-8545-fae3a15a52ca-kube-api-access-dqtgx\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.605212 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/552327a3-d3e4-4e09-8545-fae3a15a52ca-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.884310 4708 generic.go:334] "Generic (PLEG): container finished" podID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerID="98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582" exitCode=0 Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.884376 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerDied","Data":"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582"} Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.884383 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.884422 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"552327a3-d3e4-4e09-8545-fae3a15a52ca","Type":"ContainerDied","Data":"37c6b2a32ad87f14ddb29837a4a5535789e9b9b59a6806864c4498b40537ad9b"} Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.884446 4708 scope.go:117] "RemoveContainer" containerID="8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.906873 4708 scope.go:117] "RemoveContainer" containerID="110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.935873 4708 scope.go:117] "RemoveContainer" containerID="22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.942438 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.952649 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.965493 4708 scope.go:117] "RemoveContainer" containerID="98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.969336 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.969899 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="sg-core" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.969982 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="sg-core" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.970036 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-notification-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970093 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-notification-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.970182 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-central-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970228 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-central-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.970286 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="proxy-httpd" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970335 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="proxy-httpd" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970591 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="sg-core" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970654 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-notification-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970710 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="proxy-httpd" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.970765 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" containerName="ceilometer-central-agent" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.972596 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.978162 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.978307 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.978349 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.988252 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.992593 4708 scope.go:117] "RemoveContainer" containerID="8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.993332 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226\": container with ID starting with 8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226 not found: ID does not exist" containerID="8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.993385 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226"} err="failed to get container status \"8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226\": rpc error: code = NotFound desc = could not find container \"8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226\": container with ID starting with 8cc2c77717589e9b6660bd0356800edfc1bc0ad04a1d4e206f1df580a21ce226 not found: ID does not exist" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.993408 4708 scope.go:117] "RemoveContainer" containerID="110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.993861 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076\": container with ID starting with 110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076 not found: ID does not exist" containerID="110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.993884 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076"} err="failed to get container status \"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076\": rpc error: code = NotFound desc = could not find container \"110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076\": container with ID starting with 110736c999c2a25177be0282a6bedffaa9f110304c9dade5b5b054148ab54076 not found: ID does not exist" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.993900 4708 scope.go:117] "RemoveContainer" containerID="22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.994309 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938\": container with ID starting with 22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938 not found: ID does not exist" containerID="22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.994333 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938"} err="failed to get container status \"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938\": rpc error: code = NotFound desc = could not find container \"22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938\": container with ID starting with 22b7571e32bae4b4a2358013ac6d11bfe726daabb1e0a95ed70b270a62462938 not found: ID does not exist" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.994349 4708 scope.go:117] "RemoveContainer" containerID="98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582" Nov 25 05:57:43 crc kubenswrapper[4708]: E1125 05:57:43.994815 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582\": container with ID starting with 98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582 not found: ID does not exist" containerID="98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582" Nov 25 05:57:43 crc kubenswrapper[4708]: I1125 05:57:43.994837 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582"} err="failed to get container status \"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582\": rpc error: code = NotFound desc = could not find container \"98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582\": container with ID starting with 98273ec9de9ab5d8ad39ca545f840b4c3135b62adefa8090c28791198b723582 not found: ID does not exist" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014279 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-run-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014424 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-config-data\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014515 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014616 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-scripts\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014638 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-log-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014668 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x74v\" (UniqueName: \"kubernetes.io/projected/8108f47c-e809-47f5-991d-d42b19b6e8aa-kube-api-access-4x74v\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014882 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.014963 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115583 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115635 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115679 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-run-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-config-data\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115737 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115767 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-scripts\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115780 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-log-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.115797 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x74v\" (UniqueName: \"kubernetes.io/projected/8108f47c-e809-47f5-991d-d42b19b6e8aa-kube-api-access-4x74v\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.116664 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-run-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.117106 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8108f47c-e809-47f5-991d-d42b19b6e8aa-log-httpd\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.121410 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-config-data\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.126102 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.127378 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.127429 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-scripts\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.130382 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/8108f47c-e809-47f5-991d-d42b19b6e8aa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.135941 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x74v\" (UniqueName: \"kubernetes.io/projected/8108f47c-e809-47f5-991d-d42b19b6e8aa-kube-api-access-4x74v\") pod \"ceilometer-0\" (UID: \"8108f47c-e809-47f5-991d-d42b19b6e8aa\") " pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.144720 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.144731 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.288887 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.695854 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.910935 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="552327a3-d3e4-4e09-8545-fae3a15a52ca" path="/var/lib/kubelet/pods/552327a3-d3e4-4e09-8545-fae3a15a52ca/volumes" Nov 25 05:57:44 crc kubenswrapper[4708]: I1125 05:57:44.911935 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8108f47c-e809-47f5-991d-d42b19b6e8aa","Type":"ContainerStarted","Data":"0cf2e2787321e83bc39515136abeba44c4a6c7573aab2dc96bfd8bd6395dd5cd"} Nov 25 05:57:45 crc kubenswrapper[4708]: I1125 05:57:45.910000 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8108f47c-e809-47f5-991d-d42b19b6e8aa","Type":"ContainerStarted","Data":"b8eb3825a209f2c09401fab58bf971be7a109045fbb2946db7bb86c4f6a65df9"} Nov 25 05:57:45 crc kubenswrapper[4708]: I1125 05:57:45.912468 4708 generic.go:334] "Generic (PLEG): container finished" podID="cee0a911-c6e0-43b0-9b11-81f4d283dea4" containerID="961cb4a067c64ddf40f3839f15f99e2ef29c9d639cb758cc8daf5b4deb9bef7b" exitCode=0 Nov 25 05:57:45 crc kubenswrapper[4708]: I1125 05:57:45.912561 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d96zq" event={"ID":"cee0a911-c6e0-43b0-9b11-81f4d283dea4","Type":"ContainerDied","Data":"961cb4a067c64ddf40f3839f15f99e2ef29c9d639cb758cc8daf5b4deb9bef7b"} Nov 25 05:57:46 crc kubenswrapper[4708]: I1125 05:57:46.924799 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8108f47c-e809-47f5-991d-d42b19b6e8aa","Type":"ContainerStarted","Data":"0845d50613a2f798c495e9f124dcec7d2ae8a95cb304d80c63b157b4c3ed10f1"} Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.148681 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.289649 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hgcz\" (UniqueName: \"kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz\") pod \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.289707 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts\") pod \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.289813 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle\") pod \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.290716 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data\") pod \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\" (UID: \"cee0a911-c6e0-43b0-9b11-81f4d283dea4\") " Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.295634 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts" (OuterVolumeSpecName: "scripts") pod "cee0a911-c6e0-43b0-9b11-81f4d283dea4" (UID: "cee0a911-c6e0-43b0-9b11-81f4d283dea4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.296234 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz" (OuterVolumeSpecName: "kube-api-access-5hgcz") pod "cee0a911-c6e0-43b0-9b11-81f4d283dea4" (UID: "cee0a911-c6e0-43b0-9b11-81f4d283dea4"). InnerVolumeSpecName "kube-api-access-5hgcz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.318370 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cee0a911-c6e0-43b0-9b11-81f4d283dea4" (UID: "cee0a911-c6e0-43b0-9b11-81f4d283dea4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.319708 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data" (OuterVolumeSpecName: "config-data") pod "cee0a911-c6e0-43b0-9b11-81f4d283dea4" (UID: "cee0a911-c6e0-43b0-9b11-81f4d283dea4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.393023 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.393052 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hgcz\" (UniqueName: \"kubernetes.io/projected/cee0a911-c6e0-43b0-9b11-81f4d283dea4-kube-api-access-5hgcz\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.393062 4708 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.393070 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee0a911-c6e0-43b0-9b11-81f4d283dea4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.937439 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-d96zq" event={"ID":"cee0a911-c6e0-43b0-9b11-81f4d283dea4","Type":"ContainerDied","Data":"8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1"} Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.937504 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d79e6802657f7277f935e72b2435bae52ae36ac55d88bb4a83d8bab204e99c1" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.937608 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-d96zq" Nov 25 05:57:47 crc kubenswrapper[4708]: I1125 05:57:47.950032 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8108f47c-e809-47f5-991d-d42b19b6e8aa","Type":"ContainerStarted","Data":"9ddd170dba7142eca75087358bbb46fb534dfdff6314faae34d89a535d215c4b"} Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.116372 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.116641 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" containerName="nova-scheduler-scheduler" containerID="cri-o://990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d" gracePeriod=30 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.124487 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.124746 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-log" containerID="cri-o://cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516" gracePeriod=30 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.124806 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-api" containerID="cri-o://bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379" gracePeriod=30 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.223184 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.223489 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" containerID="cri-o://d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db" gracePeriod=30 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.224147 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" containerID="cri-o://3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523" gracePeriod=30 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.960253 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8108f47c-e809-47f5-991d-d42b19b6e8aa","Type":"ContainerStarted","Data":"c6fb1ca5add5005f2997b233f137f59d6f00ab1a457cfc6de45acfb099209c52"} Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.962098 4708 generic.go:334] "Generic (PLEG): container finished" podID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerID="d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db" exitCode=143 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.962119 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerDied","Data":"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db"} Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.962551 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.965125 4708 generic.go:334] "Generic (PLEG): container finished" podID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerID="cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516" exitCode=143 Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.965168 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerDied","Data":"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516"} Nov 25 05:57:48 crc kubenswrapper[4708]: I1125 05:57:48.981599 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.48073653 podStartE2EDuration="5.981588222s" podCreationTimestamp="2025-11-25 05:57:43 +0000 UTC" firstStartedPulling="2025-11-25 05:57:44.701137722 +0000 UTC m=+1006.109971109" lastFinishedPulling="2025-11-25 05:57:48.201989415 +0000 UTC m=+1009.610822801" observedRunningTime="2025-11-25 05:57:48.977880274 +0000 UTC m=+1010.386713660" watchObservedRunningTime="2025-11-25 05:57:48.981588222 +0000 UTC m=+1010.390421608" Nov 25 05:57:49 crc kubenswrapper[4708]: E1125 05:57:49.067092 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee4549fd_ebcf_4169_b8cf_04ab1e995f9f.slice/crio-34e810f1b47349b79047b9ed039dba2287d93a4a5833afe253a33452b919d925\": RecentStats: unable to find data in memory cache]" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.557713 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.643853 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle\") pod \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.644015 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfm4z\" (UniqueName: \"kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z\") pod \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.644183 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data\") pod \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\" (UID: \"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50\") " Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.650565 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z" (OuterVolumeSpecName: "kube-api-access-xfm4z") pod "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" (UID: "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50"). InnerVolumeSpecName "kube-api-access-xfm4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.671483 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" (UID: "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.672066 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data" (OuterVolumeSpecName: "config-data") pod "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" (UID: "eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.746706 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.746750 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.746763 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfm4z\" (UniqueName: \"kubernetes.io/projected/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50-kube-api-access-xfm4z\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.975588 4708 generic.go:334] "Generic (PLEG): container finished" podID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" containerID="990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d" exitCode=0 Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.975752 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50","Type":"ContainerDied","Data":"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d"} Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.976326 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50","Type":"ContainerDied","Data":"b79fe53fbfab20a3623ea556617ffbf1e3e53785b30fed3d38fcb9687bf8cbf7"} Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.976360 4708 scope.go:117] "RemoveContainer" containerID="990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d" Nov 25 05:57:49 crc kubenswrapper[4708]: I1125 05:57:49.975851 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.007933 4708 scope.go:117] "RemoveContainer" containerID="990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d" Nov 25 05:57:50 crc kubenswrapper[4708]: E1125 05:57:50.010123 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d\": container with ID starting with 990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d not found: ID does not exist" containerID="990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.010150 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d"} err="failed to get container status \"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d\": rpc error: code = NotFound desc = could not find container \"990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d\": container with ID starting with 990ea0c1a1b68d474e4028ec31b64f17543c12d6ca98804ad31a718305f75e3d not found: ID does not exist" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.019684 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.030647 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.037982 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:50 crc kubenswrapper[4708]: E1125 05:57:50.038376 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee0a911-c6e0-43b0-9b11-81f4d283dea4" containerName="nova-manage" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.038397 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee0a911-c6e0-43b0-9b11-81f4d283dea4" containerName="nova-manage" Nov 25 05:57:50 crc kubenswrapper[4708]: E1125 05:57:50.038410 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" containerName="nova-scheduler-scheduler" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.038418 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" containerName="nova-scheduler-scheduler" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.038603 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee0a911-c6e0-43b0-9b11-81f4d283dea4" containerName="nova-manage" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.038629 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" containerName="nova-scheduler-scheduler" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.039247 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.041601 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.053267 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-config-data\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.053308 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q5xqb\" (UniqueName: \"kubernetes.io/projected/815de079-46a6-4ff4-b4f3-f1e3294a21c9-kube-api-access-q5xqb\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.053411 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.057467 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.154755 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.154876 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-config-data\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.154929 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q5xqb\" (UniqueName: \"kubernetes.io/projected/815de079-46a6-4ff4-b4f3-f1e3294a21c9-kube-api-access-q5xqb\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.159372 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-config-data\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.166167 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/815de079-46a6-4ff4-b4f3-f1e3294a21c9-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.170601 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q5xqb\" (UniqueName: \"kubernetes.io/projected/815de079-46a6-4ff4-b4f3-f1e3294a21c9-kube-api-access-q5xqb\") pod \"nova-scheduler-0\" (UID: \"815de079-46a6-4ff4-b4f3-f1e3294a21c9\") " pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.361709 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.786934 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.905438 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50" path="/var/lib/kubelet/pods/eaf89dfa-bb73-4a75-b7c5-083d1fbdaa50/volumes" Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.992290 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"815de079-46a6-4ff4-b4f3-f1e3294a21c9","Type":"ContainerStarted","Data":"d76c83d74f6bb499951751592c2180c234842a491bb21223f4c589c6eddf3a10"} Nov 25 05:57:50 crc kubenswrapper[4708]: I1125 05:57:50.992645 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"815de079-46a6-4ff4-b4f3-f1e3294a21c9","Type":"ContainerStarted","Data":"fdcaa83aca8956a7df9bc421965734fc92cbef5577bd6ee963fb292fb4a404a0"} Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.012215 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.012204665 podStartE2EDuration="1.012204665s" podCreationTimestamp="2025-11-25 05:57:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:51.010370909 +0000 UTC m=+1012.419204294" watchObservedRunningTime="2025-11-25 05:57:51.012204665 +0000 UTC m=+1012.421038051" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.361448 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": read tcp 10.217.0.2:43772->10.217.0.187:8775: read: connection reset by peer" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.361498 4708 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": read tcp 10.217.0.2:43764->10.217.0.187:8775: read: connection reset by peer" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.683204 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.714066 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.795446 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.795670 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.795694 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.795758 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-szkdp\" (UniqueName: \"kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.795899 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.796013 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs\") pod \"03ec42e1-aa5d-446f-ac4a-de7559fde723\" (UID: \"03ec42e1-aa5d-446f-ac4a-de7559fde723\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.796574 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs" (OuterVolumeSpecName: "logs") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.797314 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/03ec42e1-aa5d-446f-ac4a-de7559fde723-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.801752 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp" (OuterVolumeSpecName: "kube-api-access-szkdp") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "kube-api-access-szkdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.821284 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data" (OuterVolumeSpecName: "config-data") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.821871 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.838436 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.840580 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "03ec42e1-aa5d-446f-ac4a-de7559fde723" (UID: "03ec42e1-aa5d-446f-ac4a-de7559fde723"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.898311 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle\") pod \"60c9665f-43d5-471b-9ba9-1a49b31a8435\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.898573 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs\") pod \"60c9665f-43d5-471b-9ba9-1a49b31a8435\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.898672 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs\") pod \"60c9665f-43d5-471b-9ba9-1a49b31a8435\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.898769 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfwmj\" (UniqueName: \"kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj\") pod \"60c9665f-43d5-471b-9ba9-1a49b31a8435\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.898968 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs" (OuterVolumeSpecName: "logs") pod "60c9665f-43d5-471b-9ba9-1a49b31a8435" (UID: "60c9665f-43d5-471b-9ba9-1a49b31a8435"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.899128 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data\") pod \"60c9665f-43d5-471b-9ba9-1a49b31a8435\" (UID: \"60c9665f-43d5-471b-9ba9-1a49b31a8435\") " Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900154 4708 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900179 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900192 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-szkdp\" (UniqueName: \"kubernetes.io/projected/03ec42e1-aa5d-446f-ac4a-de7559fde723-kube-api-access-szkdp\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900208 4708 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900224 4708 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/60c9665f-43d5-471b-9ba9-1a49b31a8435-logs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.900235 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ec42e1-aa5d-446f-ac4a-de7559fde723-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.903360 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj" (OuterVolumeSpecName: "kube-api-access-cfwmj") pod "60c9665f-43d5-471b-9ba9-1a49b31a8435" (UID: "60c9665f-43d5-471b-9ba9-1a49b31a8435"). InnerVolumeSpecName "kube-api-access-cfwmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.920057 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "60c9665f-43d5-471b-9ba9-1a49b31a8435" (UID: "60c9665f-43d5-471b-9ba9-1a49b31a8435"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.922360 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data" (OuterVolumeSpecName: "config-data") pod "60c9665f-43d5-471b-9ba9-1a49b31a8435" (UID: "60c9665f-43d5-471b-9ba9-1a49b31a8435"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:51 crc kubenswrapper[4708]: I1125 05:57:51.938749 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "60c9665f-43d5-471b-9ba9-1a49b31a8435" (UID: "60c9665f-43d5-471b-9ba9-1a49b31a8435"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.011223 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.011252 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.011265 4708 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/60c9665f-43d5-471b-9ba9-1a49b31a8435-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.011277 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfwmj\" (UniqueName: \"kubernetes.io/projected/60c9665f-43d5-471b-9ba9-1a49b31a8435-kube-api-access-cfwmj\") on node \"crc\" DevicePath \"\"" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.016060 4708 generic.go:334] "Generic (PLEG): container finished" podID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerID="3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523" exitCode=0 Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.016217 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerDied","Data":"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523"} Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.016314 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"60c9665f-43d5-471b-9ba9-1a49b31a8435","Type":"ContainerDied","Data":"5707dab845ce0c607b75cb12b2c2b1e546165f7313e79735846359eb3591857d"} Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.016348 4708 scope.go:117] "RemoveContainer" containerID="3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.017847 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.021170 4708 generic.go:334] "Generic (PLEG): container finished" podID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerID="bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379" exitCode=0 Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.021461 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.021610 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerDied","Data":"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379"} Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.021682 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"03ec42e1-aa5d-446f-ac4a-de7559fde723","Type":"ContainerDied","Data":"2f8cf7a077264e17c615858933545917c319b33cee3832939468b155a6471190"} Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.043845 4708 scope.go:117] "RemoveContainer" containerID="d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.068590 4708 scope.go:117] "RemoveContainer" containerID="3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.071896 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.072047 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523\": container with ID starting with 3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523 not found: ID does not exist" containerID="3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.072084 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523"} err="failed to get container status \"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523\": rpc error: code = NotFound desc = could not find container \"3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523\": container with ID starting with 3c810fa94a2c910d9eb3ccf13f4386fc09084e941f54532ebb92f862cd2d5523 not found: ID does not exist" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.072111 4708 scope.go:117] "RemoveContainer" containerID="d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.072622 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db\": container with ID starting with d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db not found: ID does not exist" containerID="d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.072649 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db"} err="failed to get container status \"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db\": rpc error: code = NotFound desc = could not find container \"d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db\": container with ID starting with d178d87b8d1137fffdddfe140f345d0f3540d14cd26f12c2e80bbccb6fd842db not found: ID does not exist" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.072669 4708 scope.go:117] "RemoveContainer" containerID="bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.091191 4708 scope.go:117] "RemoveContainer" containerID="cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.098939 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.124651 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.125113 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125133 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.125154 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-log" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125163 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-log" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.125177 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-api" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125182 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-api" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.125200 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125206 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125405 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-api" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125422 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-log" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125434 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" containerName="nova-api-log" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.125448 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" containerName="nova-metadata-metadata" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.126488 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.129261 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.129454 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.131763 4708 scope.go:117] "RemoveContainer" containerID="bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.139033 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379\": container with ID starting with bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379 not found: ID does not exist" containerID="bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.139160 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379"} err="failed to get container status \"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379\": rpc error: code = NotFound desc = could not find container \"bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379\": container with ID starting with bf4483e326b2eb5609a01b518417f9c968decfbb3efbff48f1f8b15610eab379 not found: ID does not exist" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.139187 4708 scope.go:117] "RemoveContainer" containerID="cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516" Nov 25 05:57:52 crc kubenswrapper[4708]: E1125 05:57:52.141660 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516\": container with ID starting with cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516 not found: ID does not exist" containerID="cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.141682 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516"} err="failed to get container status \"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516\": rpc error: code = NotFound desc = could not find container \"cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516\": container with ID starting with cb6f7401de0ad765948fb0f9bc35b43660309b50cf29102091f6b592a9026516 not found: ID does not exist" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.147715 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.155625 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.163917 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.169440 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.171343 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.173268 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.173365 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.173455 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.177664 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.322885 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323050 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4xxp\" (UniqueName: \"kubernetes.io/projected/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-kube-api-access-k4xxp\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323079 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323137 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-config-data\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323219 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323262 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqg2t\" (UniqueName: \"kubernetes.io/projected/12f918b9-1449-414e-abc9-723b7e31570b-kube-api-access-rqg2t\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323284 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323413 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-config-data\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323460 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323498 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-logs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.323553 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12f918b9-1449-414e-abc9-723b7e31570b-logs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425212 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-config-data\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425262 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425296 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-logs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425327 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12f918b9-1449-414e-abc9-723b7e31570b-logs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425395 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425484 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4xxp\" (UniqueName: \"kubernetes.io/projected/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-kube-api-access-k4xxp\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425512 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425597 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-config-data\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425626 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425650 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqg2t\" (UniqueName: \"kubernetes.io/projected/12f918b9-1449-414e-abc9-723b7e31570b-kube-api-access-rqg2t\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425665 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.425850 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-logs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.426579 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/12f918b9-1449-414e-abc9-723b7e31570b-logs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.430478 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.430849 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.431991 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.432027 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-config-data\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.433102 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-public-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.434667 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-config-data\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.443678 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/12f918b9-1449-414e-abc9-723b7e31570b-internal-tls-certs\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.444066 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4xxp\" (UniqueName: \"kubernetes.io/projected/c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd-kube-api-access-k4xxp\") pod \"nova-metadata-0\" (UID: \"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd\") " pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.445323 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqg2t\" (UniqueName: \"kubernetes.io/projected/12f918b9-1449-414e-abc9-723b7e31570b-kube-api-access-rqg2t\") pod \"nova-api-0\" (UID: \"12f918b9-1449-414e-abc9-723b7e31570b\") " pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.450733 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.485431 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.908163 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ec42e1-aa5d-446f-ac4a-de7559fde723" path="/var/lib/kubelet/pods/03ec42e1-aa5d-446f-ac4a-de7559fde723/volumes" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.909274 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60c9665f-43d5-471b-9ba9-1a49b31a8435" path="/var/lib/kubelet/pods/60c9665f-43d5-471b-9ba9-1a49b31a8435/volumes" Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.910771 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 05:57:52 crc kubenswrapper[4708]: W1125 05:57:52.915753 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6568a70_fb7c_4b63_9f9f_dcd3af9b47cd.slice/crio-9f408e64ed422bf3a96e858da849e9084c84223f376346c75d7b3935204c949f WatchSource:0}: Error finding container 9f408e64ed422bf3a96e858da849e9084c84223f376346c75d7b3935204c949f: Status 404 returned error can't find the container with id 9f408e64ed422bf3a96e858da849e9084c84223f376346c75d7b3935204c949f Nov 25 05:57:52 crc kubenswrapper[4708]: I1125 05:57:52.999293 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 05:57:53 crc kubenswrapper[4708]: W1125 05:57:53.014349 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12f918b9_1449_414e_abc9_723b7e31570b.slice/crio-dfaeb0dfcad428bf9a5ae5b659fb830e5c22edb2b94b4c7c2797531c06c32037 WatchSource:0}: Error finding container dfaeb0dfcad428bf9a5ae5b659fb830e5c22edb2b94b4c7c2797531c06c32037: Status 404 returned error can't find the container with id dfaeb0dfcad428bf9a5ae5b659fb830e5c22edb2b94b4c7c2797531c06c32037 Nov 25 05:57:53 crc kubenswrapper[4708]: I1125 05:57:53.036506 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12f918b9-1449-414e-abc9-723b7e31570b","Type":"ContainerStarted","Data":"dfaeb0dfcad428bf9a5ae5b659fb830e5c22edb2b94b4c7c2797531c06c32037"} Nov 25 05:57:53 crc kubenswrapper[4708]: I1125 05:57:53.040983 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd","Type":"ContainerStarted","Data":"9f408e64ed422bf3a96e858da849e9084c84223f376346c75d7b3935204c949f"} Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.069620 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd","Type":"ContainerStarted","Data":"d78ad2770543629c1172034883d0a09b8be472afca4bb636f9ece028bb15ca07"} Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.071069 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd","Type":"ContainerStarted","Data":"c1495df08cfc1e78c8796605c8f09a7754b94222c326b2dc7f8e571cfbfdef04"} Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.074714 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12f918b9-1449-414e-abc9-723b7e31570b","Type":"ContainerStarted","Data":"d205afff9fa3acc011cd6833e17f17ef19d67176f2359e0f17af9a9311c48227"} Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.074822 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"12f918b9-1449-414e-abc9-723b7e31570b","Type":"ContainerStarted","Data":"09536ae27c7a5f680f2eec8d2b7abd37551cbac043ea90ef6b80067ac5901231"} Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.093283 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.09326228 podStartE2EDuration="2.09326228s" podCreationTimestamp="2025-11-25 05:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:54.089618674 +0000 UTC m=+1015.498452059" watchObservedRunningTime="2025-11-25 05:57:54.09326228 +0000 UTC m=+1015.502095667" Nov 25 05:57:54 crc kubenswrapper[4708]: I1125 05:57:54.108090 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.10806482 podStartE2EDuration="2.10806482s" podCreationTimestamp="2025-11-25 05:57:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:57:54.105736101 +0000 UTC m=+1015.514569487" watchObservedRunningTime="2025-11-25 05:57:54.10806482 +0000 UTC m=+1015.516898207" Nov 25 05:57:55 crc kubenswrapper[4708]: I1125 05:57:55.363106 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 05:57:57 crc kubenswrapper[4708]: I1125 05:57:57.452438 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:57:57 crc kubenswrapper[4708]: I1125 05:57:57.452505 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 05:58:00 crc kubenswrapper[4708]: I1125 05:58:00.363061 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 05:58:00 crc kubenswrapper[4708]: I1125 05:58:00.388506 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 05:58:01 crc kubenswrapper[4708]: I1125 05:58:01.173270 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 05:58:02 crc kubenswrapper[4708]: I1125 05:58:02.452549 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 05:58:02 crc kubenswrapper[4708]: I1125 05:58:02.452627 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 05:58:02 crc kubenswrapper[4708]: I1125 05:58:02.487067 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:58:02 crc kubenswrapper[4708]: I1125 05:58:02.487515 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 05:58:03 crc kubenswrapper[4708]: I1125 05:58:03.468688 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:58:03 crc kubenswrapper[4708]: I1125 05:58:03.468787 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:58:03 crc kubenswrapper[4708]: I1125 05:58:03.501652 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12f918b9-1449-414e-abc9-723b7e31570b" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:58:03 crc kubenswrapper[4708]: I1125 05:58:03.501678 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="12f918b9-1449-414e-abc9-723b7e31570b" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.198:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.456825 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.457489 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.461571 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.462966 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.502597 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.503138 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.513284 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 05:58:12 crc kubenswrapper[4708]: I1125 05:58:12.525688 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 05:58:13 crc kubenswrapper[4708]: I1125 05:58:13.260987 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 05:58:13 crc kubenswrapper[4708]: I1125 05:58:13.269163 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 05:58:14 crc kubenswrapper[4708]: I1125 05:58:14.297443 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 05:58:22 crc kubenswrapper[4708]: I1125 05:58:22.292613 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:22 crc kubenswrapper[4708]: I1125 05:58:22.924941 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:26 crc kubenswrapper[4708]: I1125 05:58:26.291953 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="rabbitmq" containerID="cri-o://216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417" gracePeriod=604797 Nov 25 05:58:27 crc kubenswrapper[4708]: I1125 05:58:27.095007 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="rabbitmq" containerID="cri-o://d57993727053bb552c82679802b72966177b3586dcb69f9d5c201fd750002854" gracePeriod=604796 Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.824850 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959306 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959347 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959393 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959437 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959479 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959541 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959633 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8s4h\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959746 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959793 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959815 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959841 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf\") pod \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\" (UID: \"b9fe6604-dc04-4655-9a26-53faf3d3b6fa\") " Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.959878 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.960272 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.960922 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.962109 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.966930 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.967265 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.967762 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info" (OuterVolumeSpecName: "pod-info") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.968932 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h" (OuterVolumeSpecName: "kube-api-access-r8s4h") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "kube-api-access-r8s4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.980552 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:58:32 crc kubenswrapper[4708]: I1125 05:58:32.995060 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data" (OuterVolumeSpecName: "config-data") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064563 4708 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064627 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064640 4708 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064670 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064683 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064709 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8s4h\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-kube-api-access-r8s4h\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064719 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.064729 4708 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.067295 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf" (OuterVolumeSpecName: "server-conf") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.101984 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.102334 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.102354 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.102385 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="setup-container" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.102392 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="setup-container" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.102562 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.106176 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.108269 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.115303 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.124222 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "b9fe6604-dc04-4655-9a26-53faf3d3b6fa" (UID: "b9fe6604-dc04-4655-9a26-53faf3d3b6fa"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.133327 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166447 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166513 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166657 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166786 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9482\" (UniqueName: \"kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166811 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.166880 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.167006 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.167070 4708 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.167089 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b9fe6604-dc04-4655-9a26-53faf3d3b6fa-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.167101 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269365 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9482\" (UniqueName: \"kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269414 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269479 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269610 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269667 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269710 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.269775 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.270375 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.270588 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.270693 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.270710 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.271060 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.271315 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.284970 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9482\" (UniqueName: \"kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482\") pod \"dnsmasq-dns-74799d9859-bcxzf\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.428193 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.490189 4708 generic.go:334] "Generic (PLEG): container finished" podID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerID="d57993727053bb552c82679802b72966177b3586dcb69f9d5c201fd750002854" exitCode=0 Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.490275 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerDied","Data":"d57993727053bb552c82679802b72966177b3586dcb69f9d5c201fd750002854"} Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.493028 4708 generic.go:334] "Generic (PLEG): container finished" podID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" containerID="216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417" exitCode=0 Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.493088 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerDied","Data":"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417"} Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.493120 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b9fe6604-dc04-4655-9a26-53faf3d3b6fa","Type":"ContainerDied","Data":"aeb47cd0b95259e830debc44164147250dc98d3e28dcf9b755bf19d156b4578c"} Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.493142 4708 scope.go:117] "RemoveContainer" containerID="216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.493185 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.584418 4708 scope.go:117] "RemoveContainer" containerID="789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.631823 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.638319 4708 scope.go:117] "RemoveContainer" containerID="216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.638698 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.638771 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417\": container with ID starting with 216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417 not found: ID does not exist" containerID="216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.638815 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417"} err="failed to get container status \"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417\": rpc error: code = NotFound desc = could not find container \"216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417\": container with ID starting with 216ece983938e274082ae6ae11dbb543684eec2db3650df71cf602ca4243b417 not found: ID does not exist" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.638842 4708 scope.go:117] "RemoveContainer" containerID="789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5" Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.639114 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5\": container with ID starting with 789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5 not found: ID does not exist" containerID="789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.639148 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5"} err="failed to get container status \"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5\": rpc error: code = NotFound desc = could not find container \"789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5\": container with ID starting with 789acaef93a43c724e6a741fb5454084f0835832f7d62750a52a194e6e9815c5 not found: ID does not exist" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.653097 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.680596 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.681061 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.681081 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: E1125 05:58:33.681155 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="setup-container" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.681163 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="setup-container" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.681336 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" containerName="rabbitmq" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.682439 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.685082 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-brbzt" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.685331 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.685328 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.685536 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.685466 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.686631 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.686832 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.695922 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785221 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785278 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785315 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785373 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785431 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785488 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785530 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785593 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785645 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785758 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtkqd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.785793 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd\") pod \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\" (UID: \"2b49514f-3bad-4c22-996d-2f3c00a57ca8\") " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.786308 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.786340 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.786852 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.787163 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-config-data\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.787712 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.788087 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.788708 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74727331-8900-458f-8a20-0e8ad82d4f43-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.788783 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.788990 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.789059 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.790683 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.790816 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdfbr\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-kube-api-access-bdfbr\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.790889 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74727331-8900-458f-8a20-0e8ad82d4f43-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.790957 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.791355 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.791405 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.791420 4708 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.792481 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.792589 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.798687 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.798760 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info" (OuterVolumeSpecName: "pod-info") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.798936 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd" (OuterVolumeSpecName: "kube-api-access-rtkqd") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "kube-api-access-rtkqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.816785 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data" (OuterVolumeSpecName: "config-data") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.843315 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf" (OuterVolumeSpecName: "server-conf") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.878641 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2b49514f-3bad-4c22-996d-2f3c00a57ca8" (UID: "2b49514f-3bad-4c22-996d-2f3c00a57ca8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.897780 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74727331-8900-458f-8a20-0e8ad82d4f43-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898144 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898304 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898420 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898588 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898698 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdfbr\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-kube-api-access-bdfbr\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898604 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.898983 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.900067 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-server-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.900731 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74727331-8900-458f-8a20-0e8ad82d4f43-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.900881 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.902127 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.902266 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.902436 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-config-data\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.902860 4708 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2b49514f-3bad-4c22-996d-2f3c00a57ca8-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.903898 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.903899 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904049 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/74727331-8900-458f-8a20-0e8ad82d4f43-config-data\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.903936 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtkqd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-kube-api-access-rtkqd\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904443 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/74727331-8900-458f-8a20-0e8ad82d4f43-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904600 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904630 4708 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2b49514f-3bad-4c22-996d-2f3c00a57ca8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904665 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904681 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904693 4708 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2b49514f-3bad-4c22-996d-2f3c00a57ca8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.904703 4708 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2b49514f-3bad-4c22-996d-2f3c00a57ca8-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.908547 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.911342 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.911893 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/74727331-8900-458f-8a20-0e8ad82d4f43-pod-info\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.922475 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdfbr\" (UniqueName: \"kubernetes.io/projected/74727331-8900-458f-8a20-0e8ad82d4f43-kube-api-access-bdfbr\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.927490 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.951695 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.962270 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"74727331-8900-458f-8a20-0e8ad82d4f43\") " pod="openstack/rabbitmq-server-0" Nov 25 05:58:33 crc kubenswrapper[4708]: I1125 05:58:33.997932 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.006747 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:34 crc kubenswrapper[4708]: W1125 05:58:34.420444 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74727331_8900_458f_8a20_0e8ad82d4f43.slice/crio-c1daa900a32f54dbe7e67b45449c98fb3643448a2047e4a1222c4cae69f9eb50 WatchSource:0}: Error finding container c1daa900a32f54dbe7e67b45449c98fb3643448a2047e4a1222c4cae69f9eb50: Status 404 returned error can't find the container with id c1daa900a32f54dbe7e67b45449c98fb3643448a2047e4a1222c4cae69f9eb50 Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.422111 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.511763 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74727331-8900-458f-8a20-0e8ad82d4f43","Type":"ContainerStarted","Data":"c1daa900a32f54dbe7e67b45449c98fb3643448a2047e4a1222c4cae69f9eb50"} Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.513536 4708 generic.go:334] "Generic (PLEG): container finished" podID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerID="f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb" exitCode=0 Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.513616 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" event={"ID":"78cb9c4d-9048-4724-a29a-601cca4d6912","Type":"ContainerDied","Data":"f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb"} Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.513647 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" event={"ID":"78cb9c4d-9048-4724-a29a-601cca4d6912","Type":"ContainerStarted","Data":"43fd89c1ef5270b78ed8afc391800e53c0e40ccebb4870802281838750d77684"} Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.518383 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2b49514f-3bad-4c22-996d-2f3c00a57ca8","Type":"ContainerDied","Data":"0aab5c063b1fed19ecdce5f3bf0c14b847ba6f45acbbdf3e061db1e0830547e6"} Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.518427 4708 scope.go:117] "RemoveContainer" containerID="d57993727053bb552c82679802b72966177b3586dcb69f9d5c201fd750002854" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.518447 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.668212 4708 scope.go:117] "RemoveContainer" containerID="4a020cba17aff994f1bdc0fc2031f1ebde1a2310c8723fd78aeaf80fe4748d3b" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.690111 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.698846 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.709594 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.711614 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714036 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714076 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714177 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-l9958" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714352 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714466 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.714473 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.726867 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.727276 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833082 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833238 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833341 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833438 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66867ebe-f2fd-4936-80f8-dd70668c35ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833560 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833645 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5sb2\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-kube-api-access-n5sb2\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.833748 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.834765 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.834884 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.835020 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.835126 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66867ebe-f2fd-4936-80f8-dd70668c35ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.908703 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b49514f-3bad-4c22-996d-2f3c00a57ca8" path="/var/lib/kubelet/pods/2b49514f-3bad-4c22-996d-2f3c00a57ca8/volumes" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.910778 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9fe6604-dc04-4655-9a26-53faf3d3b6fa" path="/var/lib/kubelet/pods/b9fe6604-dc04-4655-9a26-53faf3d3b6fa/volumes" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937196 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937259 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937309 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937365 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66867ebe-f2fd-4936-80f8-dd70668c35ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937402 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937423 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937447 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937474 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66867ebe-f2fd-4936-80f8-dd70668c35ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937502 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937539 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5sb2\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-kube-api-access-n5sb2\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.937563 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.938366 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.938421 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.938584 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.938955 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.938958 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.943763 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/66867ebe-f2fd-4936-80f8-dd70668c35ab-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.944534 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.955055 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/66867ebe-f2fd-4936-80f8-dd70668c35ab-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.956033 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.958391 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/66867ebe-f2fd-4936-80f8-dd70668c35ab-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.960739 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5sb2\" (UniqueName: \"kubernetes.io/projected/66867ebe-f2fd-4936-80f8-dd70668c35ab-kube-api-access-n5sb2\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:34 crc kubenswrapper[4708]: I1125 05:58:34.968695 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"66867ebe-f2fd-4936-80f8-dd70668c35ab\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:35 crc kubenswrapper[4708]: I1125 05:58:35.047290 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:58:35 crc kubenswrapper[4708]: I1125 05:58:35.477787 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 05:58:35 crc kubenswrapper[4708]: I1125 05:58:35.539206 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" event={"ID":"78cb9c4d-9048-4724-a29a-601cca4d6912","Type":"ContainerStarted","Data":"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633"} Nov 25 05:58:35 crc kubenswrapper[4708]: I1125 05:58:35.541058 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:35 crc kubenswrapper[4708]: I1125 05:58:35.565598 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" podStartSLOduration=2.565587443 podStartE2EDuration="2.565587443s" podCreationTimestamp="2025-11-25 05:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:58:35.557537871 +0000 UTC m=+1056.966371257" watchObservedRunningTime="2025-11-25 05:58:35.565587443 +0000 UTC m=+1056.974420828" Nov 25 05:58:35 crc kubenswrapper[4708]: W1125 05:58:35.581594 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66867ebe_f2fd_4936_80f8_dd70668c35ab.slice/crio-82ffff7d410ebec87115407dbf1623413f1df5209d499f10ea4288022ff1aa3a WatchSource:0}: Error finding container 82ffff7d410ebec87115407dbf1623413f1df5209d499f10ea4288022ff1aa3a: Status 404 returned error can't find the container with id 82ffff7d410ebec87115407dbf1623413f1df5209d499f10ea4288022ff1aa3a Nov 25 05:58:36 crc kubenswrapper[4708]: I1125 05:58:36.551747 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74727331-8900-458f-8a20-0e8ad82d4f43","Type":"ContainerStarted","Data":"09cf0606b962e17e425e339e7d544cbc4ed4898b0af0d0089e4559048feadf2e"} Nov 25 05:58:36 crc kubenswrapper[4708]: I1125 05:58:36.552789 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"66867ebe-f2fd-4936-80f8-dd70668c35ab","Type":"ContainerStarted","Data":"82ffff7d410ebec87115407dbf1623413f1df5209d499f10ea4288022ff1aa3a"} Nov 25 05:58:37 crc kubenswrapper[4708]: I1125 05:58:37.566433 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"66867ebe-f2fd-4936-80f8-dd70668c35ab","Type":"ContainerStarted","Data":"f30859b3ca77ee453ab85015c356220a50cb8aee0c0c1b530f8d760077120adb"} Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.430575 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.485714 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.485955 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="dnsmasq-dns" containerID="cri-o://6c24c11eeb7b7d84c99b5fd3969b07626761c562aa0c327a656b14e0a49ab46b" gracePeriod=10 Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.593066 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d64f6bfbc-4f9hv"] Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.594770 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.612257 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d64f6bfbc-4f9hv"] Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.643503 4708 generic.go:334] "Generic (PLEG): container finished" podID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerID="6c24c11eeb7b7d84c99b5fd3969b07626761c562aa0c327a656b14e0a49ab46b" exitCode=0 Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.643645 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" event={"ID":"9aa2e102-5381-478f-91ef-8e0fe9ec13fb","Type":"ContainerDied","Data":"6c24c11eeb7b7d84c99b5fd3969b07626761c562aa0c327a656b14e0a49ab46b"} Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.744599 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.744763 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prdk9\" (UniqueName: \"kubernetes.io/projected/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-kube-api-access-prdk9\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.744845 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.744912 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.745075 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-svc\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.745145 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.745344 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-config\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.847415 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.847560 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prdk9\" (UniqueName: \"kubernetes.io/projected/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-kube-api-access-prdk9\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.847634 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.847697 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.847938 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-svc\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.848004 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.848066 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-config\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.849059 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-nb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.849121 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-svc\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.849189 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-dns-swift-storage-0\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.849376 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.849593 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-config\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.850383 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-ovsdbserver-sb\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.869083 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prdk9\" (UniqueName: \"kubernetes.io/projected/0f3ea896-7a11-4d35-94ae-0aa01889f4c8-kube-api-access-prdk9\") pod \"dnsmasq-dns-5d64f6bfbc-4f9hv\" (UID: \"0f3ea896-7a11-4d35-94ae-0aa01889f4c8\") " pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.928969 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:58:43 crc kubenswrapper[4708]: I1125 05:58:43.958862 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056249 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056375 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056501 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056596 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056691 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.056713 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dntx8\" (UniqueName: \"kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8\") pod \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\" (UID: \"9aa2e102-5381-478f-91ef-8e0fe9ec13fb\") " Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.066256 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8" (OuterVolumeSpecName: "kube-api-access-dntx8") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "kube-api-access-dntx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.112921 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config" (OuterVolumeSpecName: "config") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.117912 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.122578 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.123986 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.126450 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9aa2e102-5381-478f-91ef-8e0fe9ec13fb" (UID: "9aa2e102-5381-478f-91ef-8e0fe9ec13fb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159385 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dntx8\" (UniqueName: \"kubernetes.io/projected/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-kube-api-access-dntx8\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159417 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159429 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159438 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159447 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.159457 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9aa2e102-5381-478f-91ef-8e0fe9ec13fb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.356806 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.356873 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.373914 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d64f6bfbc-4f9hv"] Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.665415 4708 generic.go:334] "Generic (PLEG): container finished" podID="0f3ea896-7a11-4d35-94ae-0aa01889f4c8" containerID="837c04e6f636b03a298311207bf2ab592162925ac4a5074032bc58519259c279" exitCode=0 Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.665512 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" event={"ID":"0f3ea896-7a11-4d35-94ae-0aa01889f4c8","Type":"ContainerDied","Data":"837c04e6f636b03a298311207bf2ab592162925ac4a5074032bc58519259c279"} Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.665597 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" event={"ID":"0f3ea896-7a11-4d35-94ae-0aa01889f4c8","Type":"ContainerStarted","Data":"4e5089ecc922f9e1ee49dae521e8a499c06d1b5b29b032662fada1a0e0454f62"} Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.667480 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" event={"ID":"9aa2e102-5381-478f-91ef-8e0fe9ec13fb","Type":"ContainerDied","Data":"4167e595767e030e8ac4d79111a4aa1ac492ec2ed82758a9831af241bba5b2f7"} Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.667555 4708 scope.go:117] "RemoveContainer" containerID="6c24c11eeb7b7d84c99b5fd3969b07626761c562aa0c327a656b14e0a49ab46b" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.667612 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc466d6bf-tqhcq" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.764844 4708 scope.go:117] "RemoveContainer" containerID="bc97568de5f8bab32b4f9f22974bd97578b8ce69e2dd73432e7a4ccedd824a9c" Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.843304 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.848680 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc466d6bf-tqhcq"] Nov 25 05:58:44 crc kubenswrapper[4708]: I1125 05:58:44.902469 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" path="/var/lib/kubelet/pods/9aa2e102-5381-478f-91ef-8e0fe9ec13fb/volumes" Nov 25 05:58:45 crc kubenswrapper[4708]: I1125 05:58:45.680101 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" event={"ID":"0f3ea896-7a11-4d35-94ae-0aa01889f4c8","Type":"ContainerStarted","Data":"5f7221c90f12a821ddd5a48233704d92137edf15c31b2baa77b6b4bc79e676af"} Nov 25 05:58:45 crc kubenswrapper[4708]: I1125 05:58:45.680226 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:45 crc kubenswrapper[4708]: I1125 05:58:45.711409 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" podStartSLOduration=2.711386486 podStartE2EDuration="2.711386486s" podCreationTimestamp="2025-11-25 05:58:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:58:45.697983626 +0000 UTC m=+1067.106817012" watchObservedRunningTime="2025-11-25 05:58:45.711386486 +0000 UTC m=+1067.120219873" Nov 25 05:58:53 crc kubenswrapper[4708]: I1125 05:58:53.960458 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d64f6bfbc-4f9hv" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.018095 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.018350 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="dnsmasq-dns" containerID="cri-o://5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633" gracePeriod=10 Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.425861 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456124 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c9482\" (UniqueName: \"kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456193 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456236 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456310 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456384 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456407 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.456487 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config\") pod \"78cb9c4d-9048-4724-a29a-601cca4d6912\" (UID: \"78cb9c4d-9048-4724-a29a-601cca4d6912\") " Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.480352 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482" (OuterVolumeSpecName: "kube-api-access-c9482") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "kube-api-access-c9482". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.499110 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.502224 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.502925 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config" (OuterVolumeSpecName: "config") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.505299 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.509580 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.528764 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78cb9c4d-9048-4724-a29a-601cca4d6912" (UID: "78cb9c4d-9048-4724-a29a-601cca4d6912"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559743 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c9482\" (UniqueName: \"kubernetes.io/projected/78cb9c4d-9048-4724-a29a-601cca4d6912-kube-api-access-c9482\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559770 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559782 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559795 4708 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559804 4708 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559813 4708 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.559821 4708 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78cb9c4d-9048-4724-a29a-601cca4d6912-config\") on node \"crc\" DevicePath \"\"" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.784227 4708 generic.go:334] "Generic (PLEG): container finished" podID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerID="5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633" exitCode=0 Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.784280 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" event={"ID":"78cb9c4d-9048-4724-a29a-601cca4d6912","Type":"ContainerDied","Data":"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633"} Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.784315 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" event={"ID":"78cb9c4d-9048-4724-a29a-601cca4d6912","Type":"ContainerDied","Data":"43fd89c1ef5270b78ed8afc391800e53c0e40ccebb4870802281838750d77684"} Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.784324 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74799d9859-bcxzf" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.784336 4708 scope.go:117] "RemoveContainer" containerID="5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.809238 4708 scope.go:117] "RemoveContainer" containerID="f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.831313 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.840116 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74799d9859-bcxzf"] Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.841922 4708 scope.go:117] "RemoveContainer" containerID="5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633" Nov 25 05:58:54 crc kubenswrapper[4708]: E1125 05:58:54.842442 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633\": container with ID starting with 5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633 not found: ID does not exist" containerID="5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.842479 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633"} err="failed to get container status \"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633\": rpc error: code = NotFound desc = could not find container \"5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633\": container with ID starting with 5b64b623f75b4506c88f48808f9e655c1c5fa7ba0abc1f78c6ad639175554633 not found: ID does not exist" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.842511 4708 scope.go:117] "RemoveContainer" containerID="f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb" Nov 25 05:58:54 crc kubenswrapper[4708]: E1125 05:58:54.842908 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb\": container with ID starting with f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb not found: ID does not exist" containerID="f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.842952 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb"} err="failed to get container status \"f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb\": rpc error: code = NotFound desc = could not find container \"f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb\": container with ID starting with f96af6173e6ae67e003623bf7d95b3d110cdaeaf7187b7e02b1284d142def9cb not found: ID does not exist" Nov 25 05:58:54 crc kubenswrapper[4708]: I1125 05:58:54.904852 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" path="/var/lib/kubelet/pods/78cb9c4d-9048-4724-a29a-601cca4d6912/volumes" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.080966 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph"] Nov 25 05:59:06 crc kubenswrapper[4708]: E1125 05:59:06.082004 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="init" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082023 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="init" Nov 25 05:59:06 crc kubenswrapper[4708]: E1125 05:59:06.082052 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="init" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082058 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="init" Nov 25 05:59:06 crc kubenswrapper[4708]: E1125 05:59:06.082077 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082084 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: E1125 05:59:06.082096 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082101 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082329 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="78cb9c4d-9048-4724-a29a-601cca4d6912" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.082342 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aa2e102-5381-478f-91ef-8e0fe9ec13fb" containerName="dnsmasq-dns" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.083094 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.085037 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.086744 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.087542 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.092184 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph"] Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.093622 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.203844 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.204334 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.205162 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.205367 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smc5x\" (UniqueName: \"kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.307491 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.307742 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smc5x\" (UniqueName: \"kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.307901 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.308060 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.314756 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.315506 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.315809 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.324022 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smc5x\" (UniqueName: \"kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-72fph\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.406456 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.886156 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph"] Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.891169 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 05:59:06 crc kubenswrapper[4708]: I1125 05:59:06.923636 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" event={"ID":"50e3a68c-2f8a-48c7-a330-aa646ee174dd","Type":"ContainerStarted","Data":"9fa38e8db4c5cd8960d8893b173060461c4d50a0b2ca00f71674155c86f4fce9"} Nov 25 05:59:07 crc kubenswrapper[4708]: I1125 05:59:07.936398 4708 generic.go:334] "Generic (PLEG): container finished" podID="74727331-8900-458f-8a20-0e8ad82d4f43" containerID="09cf0606b962e17e425e339e7d544cbc4ed4898b0af0d0089e4559048feadf2e" exitCode=0 Nov 25 05:59:07 crc kubenswrapper[4708]: I1125 05:59:07.936503 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74727331-8900-458f-8a20-0e8ad82d4f43","Type":"ContainerDied","Data":"09cf0606b962e17e425e339e7d544cbc4ed4898b0af0d0089e4559048feadf2e"} Nov 25 05:59:08 crc kubenswrapper[4708]: I1125 05:59:08.945856 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"74727331-8900-458f-8a20-0e8ad82d4f43","Type":"ContainerStarted","Data":"057027be821c98e9b4f91246ceecb4c62f975bb643fed84a027aaf8bad678a56"} Nov 25 05:59:08 crc kubenswrapper[4708]: I1125 05:59:08.946328 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 05:59:08 crc kubenswrapper[4708]: I1125 05:59:08.947319 4708 generic.go:334] "Generic (PLEG): container finished" podID="66867ebe-f2fd-4936-80f8-dd70668c35ab" containerID="f30859b3ca77ee453ab85015c356220a50cb8aee0c0c1b530f8d760077120adb" exitCode=0 Nov 25 05:59:08 crc kubenswrapper[4708]: I1125 05:59:08.947364 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"66867ebe-f2fd-4936-80f8-dd70668c35ab","Type":"ContainerDied","Data":"f30859b3ca77ee453ab85015c356220a50cb8aee0c0c1b530f8d760077120adb"} Nov 25 05:59:08 crc kubenswrapper[4708]: I1125 05:59:08.966646 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=35.96662923 podStartE2EDuration="35.96662923s" podCreationTimestamp="2025-11-25 05:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:59:08.964013821 +0000 UTC m=+1090.372847207" watchObservedRunningTime="2025-11-25 05:59:08.96662923 +0000 UTC m=+1090.375462616" Nov 25 05:59:09 crc kubenswrapper[4708]: I1125 05:59:09.960803 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"66867ebe-f2fd-4936-80f8-dd70668c35ab","Type":"ContainerStarted","Data":"838a6ac24d82949ce75916683e9014ea210563f78aea2bbc98cdb2cb95115809"} Nov 25 05:59:09 crc kubenswrapper[4708]: I1125 05:59:09.961599 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:59:09 crc kubenswrapper[4708]: I1125 05:59:09.983079 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=35.983057978 podStartE2EDuration="35.983057978s" podCreationTimestamp="2025-11-25 05:58:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 05:59:09.978328916 +0000 UTC m=+1091.387162302" watchObservedRunningTime="2025-11-25 05:59:09.983057978 +0000 UTC m=+1091.391891364" Nov 25 05:59:14 crc kubenswrapper[4708]: I1125 05:59:14.357214 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:59:14 crc kubenswrapper[4708]: I1125 05:59:14.358028 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:59:17 crc kubenswrapper[4708]: E1125 05:59:17.223810 4708 log.go:32] "PullImage from image service failed" err="rpc error: code = Unknown desc = initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a: reading manifest sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a in quay.io/openstack-k8s-operators/openstack-ansibleee-runner: received unexpected HTTP status: 504 Gateway Time-out" image="quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a" Nov 25 05:59:17 crc kubenswrapper[4708]: E1125 05:59:17.224338 4708 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 05:59:17 crc kubenswrapper[4708]: container &Container{Name:repo-setup-edpm-deployment-openstack-edpm-ipam,Image:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a,Command:[],Args:[ansible-runner run /runner -p playbook.yaml -i repo-setup-edpm-deployment-openstack-edpm-ipam],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ANSIBLE_VERBOSITY,Value:2,ValueFrom:nil,},EnvVar{Name:RUNNER_PLAYBOOK,Value: Nov 25 05:59:17 crc kubenswrapper[4708]: - hosts: all Nov 25 05:59:17 crc kubenswrapper[4708]: strategy: linear Nov 25 05:59:17 crc kubenswrapper[4708]: tasks: Nov 25 05:59:17 crc kubenswrapper[4708]: - name: Enable podified-repos Nov 25 05:59:17 crc kubenswrapper[4708]: become: true Nov 25 05:59:17 crc kubenswrapper[4708]: ansible.builtin.shell: | Nov 25 05:59:17 crc kubenswrapper[4708]: set -euxo pipefail Nov 25 05:59:17 crc kubenswrapper[4708]: pushd /var/tmp Nov 25 05:59:17 crc kubenswrapper[4708]: curl -sL https://github.com/openstack-k8s-operators/repo-setup/archive/refs/heads/main.tar.gz | tar -xz Nov 25 05:59:17 crc kubenswrapper[4708]: pushd repo-setup-main Nov 25 05:59:17 crc kubenswrapper[4708]: python3 -m venv ./venv Nov 25 05:59:17 crc kubenswrapper[4708]: PBR_VERSION=0.0.0 ./venv/bin/pip install ./ Nov 25 05:59:17 crc kubenswrapper[4708]: ./venv/bin/repo-setup current-podified -b antelope Nov 25 05:59:17 crc kubenswrapper[4708]: popd Nov 25 05:59:17 crc kubenswrapper[4708]: rm -rf repo-setup-main Nov 25 05:59:17 crc kubenswrapper[4708]: Nov 25 05:59:17 crc kubenswrapper[4708]: Nov 25 05:59:17 crc kubenswrapper[4708]: ,ValueFrom:nil,},EnvVar{Name:RUNNER_EXTRA_VARS,Value: Nov 25 05:59:17 crc kubenswrapper[4708]: edpm_override_hosts: openstack-edpm-ipam Nov 25 05:59:17 crc kubenswrapper[4708]: edpm_service_type: repo-setup Nov 25 05:59:17 crc kubenswrapper[4708]: Nov 25 05:59:17 crc kubenswrapper[4708]: Nov 25 05:59:17 crc kubenswrapper[4708]: ,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:repo-setup-combined-ca-bundle,ReadOnly:false,MountPath:/var/lib/openstack/cacerts/repo-setup,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/runner/env/ssh_key,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:inventory,ReadOnly:false,MountPath:/runner/inventory/hosts,SubPath:inventory,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-smc5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:openstack-aee-default-env,},Optional:*true,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod repo-setup-edpm-deployment-openstack-edpm-ipam-72fph_openstack(50e3a68c-2f8a-48c7-a330-aa646ee174dd): ErrImagePull: initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a: reading manifest sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a in quay.io/openstack-k8s-operators/openstack-ansibleee-runner: received unexpected HTTP status: 504 Gateway Time-out Nov 25 05:59:17 crc kubenswrapper[4708]: > logger="UnhandledError" Nov 25 05:59:17 crc kubenswrapper[4708]: E1125 05:59:17.225517 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ErrImagePull: \"initializing source docker://quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a: reading manifest sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a in quay.io/openstack-k8s-operators/openstack-ansibleee-runner: received unexpected HTTP status: 504 Gateway Time-out\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" podUID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" Nov 25 05:59:18 crc kubenswrapper[4708]: E1125 05:59:18.035616 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"repo-setup-edpm-deployment-openstack-edpm-ipam\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:0473ff9eec0da231e2d0a10bf1abbe1dfa1a0f95b8f619e3a07605386951449a\\\"\"" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" podUID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" Nov 25 05:59:24 crc kubenswrapper[4708]: I1125 05:59:24.000740 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 05:59:25 crc kubenswrapper[4708]: I1125 05:59:25.050908 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 05:59:43 crc kubenswrapper[4708]: I1125 05:59:43.321556 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" event={"ID":"50e3a68c-2f8a-48c7-a330-aa646ee174dd","Type":"ContainerStarted","Data":"acb36effc13cf89a603395bd7c2a5a6f6d588cd611f3d65c58aa89f4e69662f6"} Nov 25 05:59:43 crc kubenswrapper[4708]: I1125 05:59:43.341853 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" podStartSLOduration=1.178689989 podStartE2EDuration="37.34183568s" podCreationTimestamp="2025-11-25 05:59:06 +0000 UTC" firstStartedPulling="2025-11-25 05:59:06.890927703 +0000 UTC m=+1088.299761088" lastFinishedPulling="2025-11-25 05:59:43.054073394 +0000 UTC m=+1124.462906779" observedRunningTime="2025-11-25 05:59:43.334972293 +0000 UTC m=+1124.743805679" watchObservedRunningTime="2025-11-25 05:59:43.34183568 +0000 UTC m=+1124.750669066" Nov 25 05:59:44 crc kubenswrapper[4708]: I1125 05:59:44.359062 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 05:59:44 crc kubenswrapper[4708]: I1125 05:59:44.359121 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 05:59:44 crc kubenswrapper[4708]: I1125 05:59:44.359166 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 05:59:44 crc kubenswrapper[4708]: I1125 05:59:44.360236 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 05:59:44 crc kubenswrapper[4708]: I1125 05:59:44.360297 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307" gracePeriod=600 Nov 25 05:59:45 crc kubenswrapper[4708]: I1125 05:59:45.352181 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307" exitCode=0 Nov 25 05:59:45 crc kubenswrapper[4708]: I1125 05:59:45.352847 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307"} Nov 25 05:59:45 crc kubenswrapper[4708]: I1125 05:59:45.352886 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527"} Nov 25 05:59:45 crc kubenswrapper[4708]: I1125 05:59:45.352910 4708 scope.go:117] "RemoveContainer" containerID="db41a0a69eac17e9745aef8980ae2454dbdf80b1197865b2abde38977096717e" Nov 25 05:59:54 crc kubenswrapper[4708]: I1125 05:59:54.457343 4708 generic.go:334] "Generic (PLEG): container finished" podID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" containerID="acb36effc13cf89a603395bd7c2a5a6f6d588cd611f3d65c58aa89f4e69662f6" exitCode=0 Nov 25 05:59:54 crc kubenswrapper[4708]: I1125 05:59:54.457394 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" event={"ID":"50e3a68c-2f8a-48c7-a330-aa646ee174dd","Type":"ContainerDied","Data":"acb36effc13cf89a603395bd7c2a5a6f6d588cd611f3d65c58aa89f4e69662f6"} Nov 25 05:59:55 crc kubenswrapper[4708]: I1125 05:59:55.906287 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.064311 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle\") pod \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.064429 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key\") pod \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.064486 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory\") pod \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.064589 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smc5x\" (UniqueName: \"kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x\") pod \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\" (UID: \"50e3a68c-2f8a-48c7-a330-aa646ee174dd\") " Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.071850 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x" (OuterVolumeSpecName: "kube-api-access-smc5x") pod "50e3a68c-2f8a-48c7-a330-aa646ee174dd" (UID: "50e3a68c-2f8a-48c7-a330-aa646ee174dd"). InnerVolumeSpecName "kube-api-access-smc5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.073334 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "50e3a68c-2f8a-48c7-a330-aa646ee174dd" (UID: "50e3a68c-2f8a-48c7-a330-aa646ee174dd"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.092136 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory" (OuterVolumeSpecName: "inventory") pod "50e3a68c-2f8a-48c7-a330-aa646ee174dd" (UID: "50e3a68c-2f8a-48c7-a330-aa646ee174dd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.092540 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "50e3a68c-2f8a-48c7-a330-aa646ee174dd" (UID: "50e3a68c-2f8a-48c7-a330-aa646ee174dd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.167124 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smc5x\" (UniqueName: \"kubernetes.io/projected/50e3a68c-2f8a-48c7-a330-aa646ee174dd-kube-api-access-smc5x\") on node \"crc\" DevicePath \"\"" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.167159 4708 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.167175 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.167193 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/50e3a68c-2f8a-48c7-a330-aa646ee174dd-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.481253 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" event={"ID":"50e3a68c-2f8a-48c7-a330-aa646ee174dd","Type":"ContainerDied","Data":"9fa38e8db4c5cd8960d8893b173060461c4d50a0b2ca00f71674155c86f4fce9"} Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.481320 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9fa38e8db4c5cd8960d8893b173060461c4d50a0b2ca00f71674155c86f4fce9" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.481356 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-72fph" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.553214 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp"] Nov 25 05:59:56 crc kubenswrapper[4708]: E1125 05:59:56.553927 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.553990 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.554235 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="50e3a68c-2f8a-48c7-a330-aa646ee174dd" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.555019 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.556756 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.557029 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.557331 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.558257 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.566134 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp"] Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.676758 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rszkz\" (UniqueName: \"kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.677100 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.677502 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.781002 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rszkz\" (UniqueName: \"kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.781096 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.781292 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.789590 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.789590 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.797197 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rszkz\" (UniqueName: \"kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-d6wtp\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:56 crc kubenswrapper[4708]: I1125 05:59:56.879765 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 05:59:57 crc kubenswrapper[4708]: I1125 05:59:57.365849 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp"] Nov 25 05:59:57 crc kubenswrapper[4708]: I1125 05:59:57.491546 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" event={"ID":"19981de6-1c24-415c-9d45-c2d31d8b54fd","Type":"ContainerStarted","Data":"fce7976fe4aa305792ded16dd8ebeaa8d677e19a9d3c4f1549d427a45d4c893f"} Nov 25 05:59:58 crc kubenswrapper[4708]: I1125 05:59:58.503104 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" event={"ID":"19981de6-1c24-415c-9d45-c2d31d8b54fd","Type":"ContainerStarted","Data":"5eae99e477d40526439e5793e683e55a5827d4c539d37ed2ca19515364656514"} Nov 25 05:59:58 crc kubenswrapper[4708]: I1125 05:59:58.536114 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" podStartSLOduration=2.029382182 podStartE2EDuration="2.536095101s" podCreationTimestamp="2025-11-25 05:59:56 +0000 UTC" firstStartedPulling="2025-11-25 05:59:57.370162003 +0000 UTC m=+1138.778995389" lastFinishedPulling="2025-11-25 05:59:57.876874933 +0000 UTC m=+1139.285708308" observedRunningTime="2025-11-25 05:59:58.528064703 +0000 UTC m=+1139.936898089" watchObservedRunningTime="2025-11-25 05:59:58.536095101 +0000 UTC m=+1139.944928487" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.136634 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc"] Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.138171 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.140061 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.140071 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.145775 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc"] Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.258392 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbvfh\" (UniqueName: \"kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.258506 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.258591 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.360576 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbvfh\" (UniqueName: \"kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.360709 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.360760 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.361797 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.367323 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.375275 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbvfh\" (UniqueName: \"kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh\") pod \"collect-profiles-29400840-w85zc\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.514778 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.522164 4708 generic.go:334] "Generic (PLEG): container finished" podID="19981de6-1c24-415c-9d45-c2d31d8b54fd" containerID="5eae99e477d40526439e5793e683e55a5827d4c539d37ed2ca19515364656514" exitCode=0 Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.522243 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" event={"ID":"19981de6-1c24-415c-9d45-c2d31d8b54fd","Type":"ContainerDied","Data":"5eae99e477d40526439e5793e683e55a5827d4c539d37ed2ca19515364656514"} Nov 25 06:00:00 crc kubenswrapper[4708]: I1125 06:00:00.920715 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc"] Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.534131 4708 generic.go:334] "Generic (PLEG): container finished" podID="3b6feecd-35f4-494c-beb1-a570abc0b382" containerID="5ea71a3d3a99497eb4cb5799a305177661bbdf49770a539a5d396e8d08d71778" exitCode=0 Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.534233 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" event={"ID":"3b6feecd-35f4-494c-beb1-a570abc0b382","Type":"ContainerDied","Data":"5ea71a3d3a99497eb4cb5799a305177661bbdf49770a539a5d396e8d08d71778"} Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.534591 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" event={"ID":"3b6feecd-35f4-494c-beb1-a570abc0b382","Type":"ContainerStarted","Data":"b322387ecfd595cc0bee9919ac6a5d7d620866709a2eaa7f3df91fb38206e408"} Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.865645 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.993154 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key\") pod \"19981de6-1c24-415c-9d45-c2d31d8b54fd\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.994055 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory\") pod \"19981de6-1c24-415c-9d45-c2d31d8b54fd\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " Nov 25 06:00:01 crc kubenswrapper[4708]: I1125 06:00:01.994202 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rszkz\" (UniqueName: \"kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz\") pod \"19981de6-1c24-415c-9d45-c2d31d8b54fd\" (UID: \"19981de6-1c24-415c-9d45-c2d31d8b54fd\") " Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.000364 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz" (OuterVolumeSpecName: "kube-api-access-rszkz") pod "19981de6-1c24-415c-9d45-c2d31d8b54fd" (UID: "19981de6-1c24-415c-9d45-c2d31d8b54fd"). InnerVolumeSpecName "kube-api-access-rszkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.021720 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "19981de6-1c24-415c-9d45-c2d31d8b54fd" (UID: "19981de6-1c24-415c-9d45-c2d31d8b54fd"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.023025 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory" (OuterVolumeSpecName: "inventory") pod "19981de6-1c24-415c-9d45-c2d31d8b54fd" (UID: "19981de6-1c24-415c-9d45-c2d31d8b54fd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.096619 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.096655 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/19981de6-1c24-415c-9d45-c2d31d8b54fd-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.096667 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rszkz\" (UniqueName: \"kubernetes.io/projected/19981de6-1c24-415c-9d45-c2d31d8b54fd-kube-api-access-rszkz\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.547265 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" event={"ID":"19981de6-1c24-415c-9d45-c2d31d8b54fd","Type":"ContainerDied","Data":"fce7976fe4aa305792ded16dd8ebeaa8d677e19a9d3c4f1549d427a45d4c893f"} Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.549249 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fce7976fe4aa305792ded16dd8ebeaa8d677e19a9d3c4f1549d427a45d4c893f" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.547313 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-d6wtp" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.605155 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj"] Nov 25 06:00:02 crc kubenswrapper[4708]: E1125 06:00:02.605549 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19981de6-1c24-415c-9d45-c2d31d8b54fd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.605569 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="19981de6-1c24-415c-9d45-c2d31d8b54fd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.605765 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="19981de6-1c24-415c-9d45-c2d31d8b54fd" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.606335 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.608120 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.608443 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.608643 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.608859 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.616163 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj"] Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.811348 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nqgf\" (UniqueName: \"kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.811690 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.811763 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.811804 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.845965 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.913450 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume\") pod \"3b6feecd-35f4-494c-beb1-a570abc0b382\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.913628 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume\") pod \"3b6feecd-35f4-494c-beb1-a570abc0b382\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.913663 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbvfh\" (UniqueName: \"kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh\") pod \"3b6feecd-35f4-494c-beb1-a570abc0b382\" (UID: \"3b6feecd-35f4-494c-beb1-a570abc0b382\") " Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.914122 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nqgf\" (UniqueName: \"kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.914147 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.914183 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.914215 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.914269 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume" (OuterVolumeSpecName: "config-volume") pod "3b6feecd-35f4-494c-beb1-a570abc0b382" (UID: "3b6feecd-35f4-494c-beb1-a570abc0b382"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.919267 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3b6feecd-35f4-494c-beb1-a570abc0b382" (UID: "3b6feecd-35f4-494c-beb1-a570abc0b382"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.919419 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh" (OuterVolumeSpecName: "kube-api-access-jbvfh") pod "3b6feecd-35f4-494c-beb1-a570abc0b382" (UID: "3b6feecd-35f4-494c-beb1-a570abc0b382"). InnerVolumeSpecName "kube-api-access-jbvfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.920005 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.920004 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.921128 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:02 crc kubenswrapper[4708]: I1125 06:00:02.927925 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nqgf\" (UniqueName: \"kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.015546 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3b6feecd-35f4-494c-beb1-a570abc0b382-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.015588 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3b6feecd-35f4-494c-beb1-a570abc0b382-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.015601 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbvfh\" (UniqueName: \"kubernetes.io/projected/3b6feecd-35f4-494c-beb1-a570abc0b382-kube-api-access-jbvfh\") on node \"crc\" DevicePath \"\"" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.220211 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.564720 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" event={"ID":"3b6feecd-35f4-494c-beb1-a570abc0b382","Type":"ContainerDied","Data":"b322387ecfd595cc0bee9919ac6a5d7d620866709a2eaa7f3df91fb38206e408"} Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.564766 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b322387ecfd595cc0bee9919ac6a5d7d620866709a2eaa7f3df91fb38206e408" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.564775 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc" Nov 25 06:00:03 crc kubenswrapper[4708]: I1125 06:00:03.773578 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj"] Nov 25 06:00:04 crc kubenswrapper[4708]: I1125 06:00:04.580154 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" event={"ID":"eecde679-d675-471a-9e80-a38eca4feba7","Type":"ContainerStarted","Data":"7669b3f6af58eb43514ebe86ee6fd91c6a6320a70d2e3c07c06a68861d75dd31"} Nov 25 06:00:04 crc kubenswrapper[4708]: I1125 06:00:04.580217 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" event={"ID":"eecde679-d675-471a-9e80-a38eca4feba7","Type":"ContainerStarted","Data":"f8c335db93411a33dad704b9f3a45eb29ae6a68e0e8afec828be9c79f91c91cf"} Nov 25 06:00:04 crc kubenswrapper[4708]: I1125 06:00:04.601185 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" podStartSLOduration=2.097063961 podStartE2EDuration="2.601172421s" podCreationTimestamp="2025-11-25 06:00:02 +0000 UTC" firstStartedPulling="2025-11-25 06:00:03.762643239 +0000 UTC m=+1145.171476625" lastFinishedPulling="2025-11-25 06:00:04.266751699 +0000 UTC m=+1145.675585085" observedRunningTime="2025-11-25 06:00:04.594180792 +0000 UTC m=+1146.003014178" watchObservedRunningTime="2025-11-25 06:00:04.601172421 +0000 UTC m=+1146.010005808" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.134766 4708 scope.go:117] "RemoveContainer" containerID="c4b97ef4e5af2850fbd1c6e7912f5404dded7fd308785bf0b12be940cd5ffa35" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.141160 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29400841-fnx8l"] Nov 25 06:01:00 crc kubenswrapper[4708]: E1125 06:01:00.141758 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b6feecd-35f4-494c-beb1-a570abc0b382" containerName="collect-profiles" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.141781 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b6feecd-35f4-494c-beb1-a570abc0b382" containerName="collect-profiles" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.141961 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b6feecd-35f4-494c-beb1-a570abc0b382" containerName="collect-profiles" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.142600 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.152404 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400841-fnx8l"] Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.164774 4708 scope.go:117] "RemoveContainer" containerID="72592f690d89e2a157c7daf0ef72a2c25c4048de91cd5c62518f2fba7db4e947" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.192843 4708 scope.go:117] "RemoveContainer" containerID="abbafe6a07a038f715d82b1a8caf876dcca2d534627ad779c64b17e6e5ac4455" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.235027 4708 scope.go:117] "RemoveContainer" containerID="f960cc33ea47dc79cf61a83814f0b8f718685a43ff98f0acaa071283b85d9daf" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.266742 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltnp4\" (UniqueName: \"kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.266848 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.268717 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.268843 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.371618 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.371818 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.371901 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.371961 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltnp4\" (UniqueName: \"kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.378662 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.378679 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.379724 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.387931 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltnp4\" (UniqueName: \"kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4\") pod \"keystone-cron-29400841-fnx8l\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.461281 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:00 crc kubenswrapper[4708]: I1125 06:01:00.868334 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29400841-fnx8l"] Nov 25 06:01:01 crc kubenswrapper[4708]: I1125 06:01:01.164135 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400841-fnx8l" event={"ID":"92abfbe3-d658-4e40-aa2b-38167fe164fd","Type":"ContainerStarted","Data":"4d6ddb859e9840e8724dbb886a3feff3114c7ddaa1d71745a3b322a5fbbe4b24"} Nov 25 06:01:01 crc kubenswrapper[4708]: I1125 06:01:01.164468 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400841-fnx8l" event={"ID":"92abfbe3-d658-4e40-aa2b-38167fe164fd","Type":"ContainerStarted","Data":"22daf40d76dcdb13ac8d9984ff8eea25695a59385389e0c52e3dbf5f9fc45554"} Nov 25 06:01:01 crc kubenswrapper[4708]: I1125 06:01:01.194538 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29400841-fnx8l" podStartSLOduration=1.194506286 podStartE2EDuration="1.194506286s" podCreationTimestamp="2025-11-25 06:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 06:01:01.184366603 +0000 UTC m=+1202.593199989" watchObservedRunningTime="2025-11-25 06:01:01.194506286 +0000 UTC m=+1202.603339673" Nov 25 06:01:03 crc kubenswrapper[4708]: E1125 06:01:03.472576 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92abfbe3_d658_4e40_aa2b_38167fe164fd.slice/crio-conmon-4d6ddb859e9840e8724dbb886a3feff3114c7ddaa1d71745a3b322a5fbbe4b24.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92abfbe3_d658_4e40_aa2b_38167fe164fd.slice/crio-4d6ddb859e9840e8724dbb886a3feff3114c7ddaa1d71745a3b322a5fbbe4b24.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:01:04 crc kubenswrapper[4708]: I1125 06:01:04.192953 4708 generic.go:334] "Generic (PLEG): container finished" podID="92abfbe3-d658-4e40-aa2b-38167fe164fd" containerID="4d6ddb859e9840e8724dbb886a3feff3114c7ddaa1d71745a3b322a5fbbe4b24" exitCode=0 Nov 25 06:01:04 crc kubenswrapper[4708]: I1125 06:01:04.193028 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400841-fnx8l" event={"ID":"92abfbe3-d658-4e40-aa2b-38167fe164fd","Type":"ContainerDied","Data":"4d6ddb859e9840e8724dbb886a3feff3114c7ddaa1d71745a3b322a5fbbe4b24"} Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.511245 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.588546 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data\") pod \"92abfbe3-d658-4e40-aa2b-38167fe164fd\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.588670 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle\") pod \"92abfbe3-d658-4e40-aa2b-38167fe164fd\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.588716 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltnp4\" (UniqueName: \"kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4\") pod \"92abfbe3-d658-4e40-aa2b-38167fe164fd\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.588830 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys\") pod \"92abfbe3-d658-4e40-aa2b-38167fe164fd\" (UID: \"92abfbe3-d658-4e40-aa2b-38167fe164fd\") " Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.595516 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "92abfbe3-d658-4e40-aa2b-38167fe164fd" (UID: "92abfbe3-d658-4e40-aa2b-38167fe164fd"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.597329 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4" (OuterVolumeSpecName: "kube-api-access-ltnp4") pod "92abfbe3-d658-4e40-aa2b-38167fe164fd" (UID: "92abfbe3-d658-4e40-aa2b-38167fe164fd"). InnerVolumeSpecName "kube-api-access-ltnp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.616278 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92abfbe3-d658-4e40-aa2b-38167fe164fd" (UID: "92abfbe3-d658-4e40-aa2b-38167fe164fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.640706 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data" (OuterVolumeSpecName: "config-data") pod "92abfbe3-d658-4e40-aa2b-38167fe164fd" (UID: "92abfbe3-d658-4e40-aa2b-38167fe164fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.691107 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.691142 4708 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.691156 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltnp4\" (UniqueName: \"kubernetes.io/projected/92abfbe3-d658-4e40-aa2b-38167fe164fd-kube-api-access-ltnp4\") on node \"crc\" DevicePath \"\"" Nov 25 06:01:05 crc kubenswrapper[4708]: I1125 06:01:05.691169 4708 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92abfbe3-d658-4e40-aa2b-38167fe164fd-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 06:01:06 crc kubenswrapper[4708]: I1125 06:01:06.226537 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29400841-fnx8l" event={"ID":"92abfbe3-d658-4e40-aa2b-38167fe164fd","Type":"ContainerDied","Data":"22daf40d76dcdb13ac8d9984ff8eea25695a59385389e0c52e3dbf5f9fc45554"} Nov 25 06:01:06 crc kubenswrapper[4708]: I1125 06:01:06.226602 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22daf40d76dcdb13ac8d9984ff8eea25695a59385389e0c52e3dbf5f9fc45554" Nov 25 06:01:06 crc kubenswrapper[4708]: I1125 06:01:06.226653 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29400841-fnx8l" Nov 25 06:01:44 crc kubenswrapper[4708]: I1125 06:01:44.356717 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:01:44 crc kubenswrapper[4708]: I1125 06:01:44.357383 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:02:14 crc kubenswrapper[4708]: I1125 06:02:14.357590 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:02:14 crc kubenswrapper[4708]: I1125 06:02:14.358356 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:02:44 crc kubenswrapper[4708]: I1125 06:02:44.356931 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:02:44 crc kubenswrapper[4708]: I1125 06:02:44.357647 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:02:44 crc kubenswrapper[4708]: I1125 06:02:44.357700 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:02:44 crc kubenswrapper[4708]: I1125 06:02:44.358923 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:02:44 crc kubenswrapper[4708]: I1125 06:02:44.358982 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527" gracePeriod=600 Nov 25 06:02:45 crc kubenswrapper[4708]: I1125 06:02:45.155107 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527" exitCode=0 Nov 25 06:02:45 crc kubenswrapper[4708]: I1125 06:02:45.155167 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527"} Nov 25 06:02:45 crc kubenswrapper[4708]: I1125 06:02:45.155747 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf"} Nov 25 06:02:45 crc kubenswrapper[4708]: I1125 06:02:45.155774 4708 scope.go:117] "RemoveContainer" containerID="f83ab71ea3b981acaba6e7bb989ecc0cb85a2e318c0fcd05bdcc2f2ccaf8c307" Nov 25 06:03:22 crc kubenswrapper[4708]: I1125 06:03:22.495482 4708 generic.go:334] "Generic (PLEG): container finished" podID="eecde679-d675-471a-9e80-a38eca4feba7" containerID="7669b3f6af58eb43514ebe86ee6fd91c6a6320a70d2e3c07c06a68861d75dd31" exitCode=0 Nov 25 06:03:22 crc kubenswrapper[4708]: I1125 06:03:22.495589 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" event={"ID":"eecde679-d675-471a-9e80-a38eca4feba7","Type":"ContainerDied","Data":"7669b3f6af58eb43514ebe86ee6fd91c6a6320a70d2e3c07c06a68861d75dd31"} Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.820238 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.955553 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5nqgf\" (UniqueName: \"kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf\") pod \"eecde679-d675-471a-9e80-a38eca4feba7\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.955738 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory\") pod \"eecde679-d675-471a-9e80-a38eca4feba7\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.955855 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle\") pod \"eecde679-d675-471a-9e80-a38eca4feba7\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.956054 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key\") pod \"eecde679-d675-471a-9e80-a38eca4feba7\" (UID: \"eecde679-d675-471a-9e80-a38eca4feba7\") " Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.964182 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf" (OuterVolumeSpecName: "kube-api-access-5nqgf") pod "eecde679-d675-471a-9e80-a38eca4feba7" (UID: "eecde679-d675-471a-9e80-a38eca4feba7"). InnerVolumeSpecName "kube-api-access-5nqgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.964317 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "eecde679-d675-471a-9e80-a38eca4feba7" (UID: "eecde679-d675-471a-9e80-a38eca4feba7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.986328 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eecde679-d675-471a-9e80-a38eca4feba7" (UID: "eecde679-d675-471a-9e80-a38eca4feba7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:03:23 crc kubenswrapper[4708]: I1125 06:03:23.986588 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory" (OuterVolumeSpecName: "inventory") pod "eecde679-d675-471a-9e80-a38eca4feba7" (UID: "eecde679-d675-471a-9e80-a38eca4feba7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.057716 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.057749 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5nqgf\" (UniqueName: \"kubernetes.io/projected/eecde679-d675-471a-9e80-a38eca4feba7-kube-api-access-5nqgf\") on node \"crc\" DevicePath \"\"" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.057762 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.057774 4708 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eecde679-d675-471a-9e80-a38eca4feba7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.516849 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" event={"ID":"eecde679-d675-471a-9e80-a38eca4feba7","Type":"ContainerDied","Data":"f8c335db93411a33dad704b9f3a45eb29ae6a68e0e8afec828be9c79f91c91cf"} Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.516911 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8c335db93411a33dad704b9f3a45eb29ae6a68e0e8afec828be9c79f91c91cf" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.516920 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.578976 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm"] Nov 25 06:03:24 crc kubenswrapper[4708]: E1125 06:03:24.579795 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eecde679-d675-471a-9e80-a38eca4feba7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.579817 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="eecde679-d675-471a-9e80-a38eca4feba7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 06:03:24 crc kubenswrapper[4708]: E1125 06:03:24.579834 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92abfbe3-d658-4e40-aa2b-38167fe164fd" containerName="keystone-cron" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.579841 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="92abfbe3-d658-4e40-aa2b-38167fe164fd" containerName="keystone-cron" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.580021 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="eecde679-d675-471a-9e80-a38eca4feba7" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.580052 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="92abfbe3-d658-4e40-aa2b-38167fe164fd" containerName="keystone-cron" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.580692 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.584881 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.584909 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.584988 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.585324 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.602780 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm"] Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.771952 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.772009 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.772106 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7cn42\" (UniqueName: \"kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.873561 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.873624 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.873667 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7cn42\" (UniqueName: \"kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.878314 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.878545 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.891600 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7cn42\" (UniqueName: \"kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:24 crc kubenswrapper[4708]: I1125 06:03:24.900149 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:03:25 crc kubenswrapper[4708]: I1125 06:03:25.370818 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm"] Nov 25 06:03:25 crc kubenswrapper[4708]: I1125 06:03:25.527647 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" event={"ID":"555ae52e-3696-4f22-b959-d5b9407242bf","Type":"ContainerStarted","Data":"72e816de084964c72acb9983f1c4f8980a9f6640d60cb2eb3d48ae4d7d8f2b45"} Nov 25 06:03:26 crc kubenswrapper[4708]: I1125 06:03:26.537165 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" event={"ID":"555ae52e-3696-4f22-b959-d5b9407242bf","Type":"ContainerStarted","Data":"fdde5b65f3f928d789db19f50f4bab4859eaa4e09ff92ff3e240b432edec5bb5"} Nov 25 06:03:26 crc kubenswrapper[4708]: I1125 06:03:26.562387 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" podStartSLOduration=2.038185799 podStartE2EDuration="2.562360783s" podCreationTimestamp="2025-11-25 06:03:24 +0000 UTC" firstStartedPulling="2025-11-25 06:03:25.377579477 +0000 UTC m=+1346.786412863" lastFinishedPulling="2025-11-25 06:03:25.901754461 +0000 UTC m=+1347.310587847" observedRunningTime="2025-11-25 06:03:26.556809206 +0000 UTC m=+1347.965642592" watchObservedRunningTime="2025-11-25 06:03:26.562360783 +0000 UTC m=+1347.971194169" Nov 25 06:04:11 crc kubenswrapper[4708]: I1125 06:04:11.038991 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-57c9k"] Nov 25 06:04:11 crc kubenswrapper[4708]: I1125 06:04:11.045083 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-fb32-account-create-5t5ts"] Nov 25 06:04:11 crc kubenswrapper[4708]: I1125 06:04:11.050509 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-57c9k"] Nov 25 06:04:11 crc kubenswrapper[4708]: I1125 06:04:11.062746 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-fb32-account-create-5t5ts"] Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.038383 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-sxdxd"] Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.044130 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6ca0-account-create-x9lqj"] Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.052304 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6ca0-account-create-x9lqj"] Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.065864 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-sxdxd"] Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.905727 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2076e535-3cc1-49be-8002-cb063502685c" path="/var/lib/kubelet/pods/2076e535-3cc1-49be-8002-cb063502685c/volumes" Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.906946 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28" path="/var/lib/kubelet/pods/8ffa53bf-ef4e-4d48-953e-bfcdb8e8bd28/volumes" Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.907552 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c840b75e-0e64-448f-ba51-051c712afe4d" path="/var/lib/kubelet/pods/c840b75e-0e64-448f-ba51-051c712afe4d/volumes" Nov 25 06:04:12 crc kubenswrapper[4708]: I1125 06:04:12.908112 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce71c684-746d-4e4e-ad21-2f80340fd52b" path="/var/lib/kubelet/pods/ce71c684-746d-4e4e-ad21-2f80340fd52b/volumes" Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.033920 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-5c67-account-create-k6fhh"] Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.042685 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-5c67-account-create-k6fhh"] Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.049991 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-n4fvw"] Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.054501 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-n4fvw"] Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.902588 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cb2b1fd-09b1-476b-a6ad-ae22500b5467" path="/var/lib/kubelet/pods/2cb2b1fd-09b1-476b-a6ad-ae22500b5467/volumes" Nov 25 06:04:18 crc kubenswrapper[4708]: I1125 06:04:18.903187 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4011a3c5-2e4b-410b-88c6-dfc229b330c2" path="/var/lib/kubelet/pods/4011a3c5-2e4b-410b-88c6-dfc229b330c2/volumes" Nov 25 06:04:44 crc kubenswrapper[4708]: I1125 06:04:44.357254 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:04:44 crc kubenswrapper[4708]: I1125 06:04:44.357830 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.034542 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-6f71-account-create-65cgg"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.041739 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-vw8x4"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.050753 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f6ba-account-create-2c6rf"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.058535 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-6f71-account-create-65cgg"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.064692 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-b25rc"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.069904 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-5hcbt"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.074903 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-b25rc"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.080152 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-vw8x4"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.085224 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-a6c7-account-create-z2ldb"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.090353 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f6ba-account-create-2c6rf"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.101609 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-5hcbt"] Nov 25 06:04:45 crc kubenswrapper[4708]: I1125 06:04:45.106195 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-a6c7-account-create-z2ldb"] Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.906840 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08c33523-d685-4072-9673-9989ed7ad578" path="/var/lib/kubelet/pods/08c33523-d685-4072-9673-9989ed7ad578/volumes" Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.907879 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23de5970-5eb9-4f95-b155-e7bc2fdbed7a" path="/var/lib/kubelet/pods/23de5970-5eb9-4f95-b155-e7bc2fdbed7a/volumes" Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.908417 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79753d96-259b-4fd5-9dfe-cb27f6ce82d6" path="/var/lib/kubelet/pods/79753d96-259b-4fd5-9dfe-cb27f6ce82d6/volumes" Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.908955 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c883a5b-0f2a-48a0-b29c-9bfc01fb439d" path="/var/lib/kubelet/pods/7c883a5b-0f2a-48a0-b29c-9bfc01fb439d/volumes" Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.910001 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c8aaf81-e916-4011-be02-a76c72a956f8" path="/var/lib/kubelet/pods/9c8aaf81-e916-4011-be02-a76c72a956f8/volumes" Nov 25 06:04:46 crc kubenswrapper[4708]: I1125 06:04:46.910560 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a40cd066-4de2-427b-b1f6-207271f43d26" path="/var/lib/kubelet/pods/a40cd066-4de2-427b-b1f6-207271f43d26/volumes" Nov 25 06:04:50 crc kubenswrapper[4708]: I1125 06:04:50.033068 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-qv9pr"] Nov 25 06:04:50 crc kubenswrapper[4708]: I1125 06:04:50.043970 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-qv9pr"] Nov 25 06:04:50 crc kubenswrapper[4708]: I1125 06:04:50.903480 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc02d86b-b016-4e10-960e-f7237575e10d" path="/var/lib/kubelet/pods/cc02d86b-b016-4e10-960e-f7237575e10d/volumes" Nov 25 06:04:54 crc kubenswrapper[4708]: I1125 06:04:54.031560 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wxksw"] Nov 25 06:04:54 crc kubenswrapper[4708]: I1125 06:04:54.037619 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wxksw"] Nov 25 06:04:54 crc kubenswrapper[4708]: I1125 06:04:54.904301 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f8dae9-be16-4ee1-a485-73148c832c47" path="/var/lib/kubelet/pods/31f8dae9-be16-4ee1-a485-73148c832c47/volumes" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.382157 4708 scope.go:117] "RemoveContainer" containerID="f3bce6e1deb8f06f80c37c18d8ed37882c3b2c6e7a42e95c8a043f3e0366e64d" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.410646 4708 scope.go:117] "RemoveContainer" containerID="7e45c73f9d401a7dfc1dfdcb96333776b6974785571bd02b3e9a28860047f8c1" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.438635 4708 scope.go:117] "RemoveContainer" containerID="2224ff2439a37fecc13859d4d06accb76f8390714d2ebd25431b1a9ceb639735" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.480357 4708 scope.go:117] "RemoveContainer" containerID="1d45c97abf29d35ddc72ef5becca9092ddd639cb53c5ca73285d0029fb1f66f0" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.516799 4708 scope.go:117] "RemoveContainer" containerID="eefe49bf32dbbf9245728e7fe4d324297b299f1dfb6537d8688c8f8fabee86d6" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.534995 4708 scope.go:117] "RemoveContainer" containerID="3bf296dc6137b503c41f0b1816414fee658b22f92c62a442725387b7fc8a7e56" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.573657 4708 scope.go:117] "RemoveContainer" containerID="476d5209b4027ed7175a8cd5fc91f95fad8a6285d08e6ca682c529e872d8e508" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.591564 4708 scope.go:117] "RemoveContainer" containerID="7a0ab90e47fcbdb851207e60b1ad0b1bfe91ee1679d8de0431795195bb4eabc5" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.608884 4708 scope.go:117] "RemoveContainer" containerID="79113f94b45daa9f043c856c87c0a8bae146cf04c735e7d4eba70a57ff57bd89" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.623346 4708 scope.go:117] "RemoveContainer" containerID="0a59cfb5eff3cd4addb0fa6cbd7300e17102ee45f5f7324b63f61f6d07bc2b05" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.638500 4708 scope.go:117] "RemoveContainer" containerID="992d08eea81ff9987f9e2d771426ee188c50312709dc0d38f1e07503bfadf29b" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.652451 4708 scope.go:117] "RemoveContainer" containerID="2e714e36b81afda9a8e91f3780d95e2c3137f3c83f2834e7304dff6f43a348f7" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.674738 4708 scope.go:117] "RemoveContainer" containerID="575e8ec457441dfba1ae5262caec6f0b82e73b2fcb797c1aca635e06565edce9" Nov 25 06:05:00 crc kubenswrapper[4708]: I1125 06:05:00.693169 4708 scope.go:117] "RemoveContainer" containerID="0ac1c5cf865fd6a4fe3a584bbf1264f52ab177988d9311eb3e68a66a79205eeb" Nov 25 06:05:07 crc kubenswrapper[4708]: I1125 06:05:07.492316 4708 generic.go:334] "Generic (PLEG): container finished" podID="555ae52e-3696-4f22-b959-d5b9407242bf" containerID="fdde5b65f3f928d789db19f50f4bab4859eaa4e09ff92ff3e240b432edec5bb5" exitCode=0 Nov 25 06:05:07 crc kubenswrapper[4708]: I1125 06:05:07.492408 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" event={"ID":"555ae52e-3696-4f22-b959-d5b9407242bf","Type":"ContainerDied","Data":"fdde5b65f3f928d789db19f50f4bab4859eaa4e09ff92ff3e240b432edec5bb5"} Nov 25 06:05:08 crc kubenswrapper[4708]: I1125 06:05:08.807149 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:05:08 crc kubenswrapper[4708]: I1125 06:05:08.991536 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory\") pod \"555ae52e-3696-4f22-b959-d5b9407242bf\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " Nov 25 06:05:08 crc kubenswrapper[4708]: I1125 06:05:08.991808 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7cn42\" (UniqueName: \"kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42\") pod \"555ae52e-3696-4f22-b959-d5b9407242bf\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " Nov 25 06:05:08 crc kubenswrapper[4708]: I1125 06:05:08.991876 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key\") pod \"555ae52e-3696-4f22-b959-d5b9407242bf\" (UID: \"555ae52e-3696-4f22-b959-d5b9407242bf\") " Nov 25 06:05:08 crc kubenswrapper[4708]: I1125 06:05:08.997365 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42" (OuterVolumeSpecName: "kube-api-access-7cn42") pod "555ae52e-3696-4f22-b959-d5b9407242bf" (UID: "555ae52e-3696-4f22-b959-d5b9407242bf"). InnerVolumeSpecName "kube-api-access-7cn42". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.019184 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory" (OuterVolumeSpecName: "inventory") pod "555ae52e-3696-4f22-b959-d5b9407242bf" (UID: "555ae52e-3696-4f22-b959-d5b9407242bf"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.020386 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "555ae52e-3696-4f22-b959-d5b9407242bf" (UID: "555ae52e-3696-4f22-b959-d5b9407242bf"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.044029 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-vcmh6"] Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.050740 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-vcmh6"] Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.095088 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.095371 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/555ae52e-3696-4f22-b959-d5b9407242bf-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.095385 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7cn42\" (UniqueName: \"kubernetes.io/projected/555ae52e-3696-4f22-b959-d5b9407242bf-kube-api-access-7cn42\") on node \"crc\" DevicePath \"\"" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.512440 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" event={"ID":"555ae52e-3696-4f22-b959-d5b9407242bf","Type":"ContainerDied","Data":"72e816de084964c72acb9983f1c4f8980a9f6640d60cb2eb3d48ae4d7d8f2b45"} Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.512496 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72e816de084964c72acb9983f1c4f8980a9f6640d60cb2eb3d48ae4d7d8f2b45" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.512512 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.571943 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl"] Nov 25 06:05:09 crc kubenswrapper[4708]: E1125 06:05:09.572467 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555ae52e-3696-4f22-b959-d5b9407242bf" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.572490 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="555ae52e-3696-4f22-b959-d5b9407242bf" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.572693 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="555ae52e-3696-4f22-b959-d5b9407242bf" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.573339 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.575783 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.578476 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.578692 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.578884 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.580386 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl"] Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.706622 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.706904 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.707027 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwx5d\" (UniqueName: \"kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.808295 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.808377 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.808423 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwx5d\" (UniqueName: \"kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.813323 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.814331 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.822901 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwx5d\" (UniqueName: \"kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:09 crc kubenswrapper[4708]: I1125 06:05:09.889149 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:05:10 crc kubenswrapper[4708]: I1125 06:05:10.346864 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl"] Nov 25 06:05:10 crc kubenswrapper[4708]: I1125 06:05:10.348422 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:05:10 crc kubenswrapper[4708]: I1125 06:05:10.520968 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" event={"ID":"87296b4d-6e9f-4206-9d0f-78dd40e51e17","Type":"ContainerStarted","Data":"e490b4bf5caadeb34f9d51c400cdfe8a1b2a5ce7ad57c30f787b7fad83eb0ef9"} Nov 25 06:05:10 crc kubenswrapper[4708]: I1125 06:05:10.906228 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9174a689-38fd-478f-8c3e-07ab7e21a921" path="/var/lib/kubelet/pods/9174a689-38fd-478f-8c3e-07ab7e21a921/volumes" Nov 25 06:05:11 crc kubenswrapper[4708]: I1125 06:05:11.532009 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" event={"ID":"87296b4d-6e9f-4206-9d0f-78dd40e51e17","Type":"ContainerStarted","Data":"579560a0e9a1e30cf10588d2bada08120ec252aacb6534a285afea854cfb4b8b"} Nov 25 06:05:11 crc kubenswrapper[4708]: I1125 06:05:11.547945 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" podStartSLOduration=1.973104164 podStartE2EDuration="2.547925142s" podCreationTimestamp="2025-11-25 06:05:09 +0000 UTC" firstStartedPulling="2025-11-25 06:05:10.348215872 +0000 UTC m=+1451.757049259" lastFinishedPulling="2025-11-25 06:05:10.923036851 +0000 UTC m=+1452.331870237" observedRunningTime="2025-11-25 06:05:11.543759791 +0000 UTC m=+1452.952593177" watchObservedRunningTime="2025-11-25 06:05:11.547925142 +0000 UTC m=+1452.956758529" Nov 25 06:05:14 crc kubenswrapper[4708]: I1125 06:05:14.357464 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:05:14 crc kubenswrapper[4708]: I1125 06:05:14.357878 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.027377 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cqjrf"] Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.033960 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-lgvsg"] Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.039455 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cqjrf"] Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.044413 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-lgvsg"] Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.903070 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f0e677f-bc35-4c02-8322-37d4cf7fc350" path="/var/lib/kubelet/pods/0f0e677f-bc35-4c02-8322-37d4cf7fc350/volumes" Nov 25 06:05:24 crc kubenswrapper[4708]: I1125 06:05:24.904375 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5726473d-059a-4f7d-ac6b-4762c732f8ff" path="/var/lib/kubelet/pods/5726473d-059a-4f7d-ac6b-4762c732f8ff/volumes" Nov 25 06:05:28 crc kubenswrapper[4708]: I1125 06:05:28.028966 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8kfbn"] Nov 25 06:05:28 crc kubenswrapper[4708]: I1125 06:05:28.036229 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8kfbn"] Nov 25 06:05:28 crc kubenswrapper[4708]: I1125 06:05:28.903125 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75e750a1-381d-4d38-8d4b-75ac9b3ea2d4" path="/var/lib/kubelet/pods/75e750a1-381d-4d38-8d4b-75ac9b3ea2d4/volumes" Nov 25 06:05:30 crc kubenswrapper[4708]: I1125 06:05:30.030146 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-gnqch"] Nov 25 06:05:30 crc kubenswrapper[4708]: I1125 06:05:30.037583 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-gnqch"] Nov 25 06:05:30 crc kubenswrapper[4708]: I1125 06:05:30.903152 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f" path="/var/lib/kubelet/pods/e0e9c3df-0e3e-4c24-bf30-7d9bb11d330f/volumes" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.356675 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.357354 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.357408 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.358802 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.358865 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" gracePeriod=600 Nov 25 06:05:44 crc kubenswrapper[4708]: E1125 06:05:44.481320 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.825125 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" exitCode=0 Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.825199 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf"} Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.825280 4708 scope.go:117] "RemoveContainer" containerID="9b9861b396058178e4d76a8f183b8cf99f2760f02601b2c91fe1d3788d338527" Nov 25 06:05:44 crc kubenswrapper[4708]: I1125 06:05:44.826119 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:05:44 crc kubenswrapper[4708]: E1125 06:05:44.826428 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.474797 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.477156 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.488037 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.520035 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.520210 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdfgt\" (UniqueName: \"kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.520296 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.622720 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.622810 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdfgt\" (UniqueName: \"kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.622851 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.623289 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.623620 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.640881 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdfgt\" (UniqueName: \"kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt\") pod \"redhat-operators-smlw2\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:51 crc kubenswrapper[4708]: I1125 06:05:51.793952 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:05:52 crc kubenswrapper[4708]: I1125 06:05:52.212139 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:05:52 crc kubenswrapper[4708]: I1125 06:05:52.911958 4708 generic.go:334] "Generic (PLEG): container finished" podID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerID="ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97" exitCode=0 Nov 25 06:05:52 crc kubenswrapper[4708]: I1125 06:05:52.912101 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerDied","Data":"ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97"} Nov 25 06:05:52 crc kubenswrapper[4708]: I1125 06:05:52.912345 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerStarted","Data":"ecbde0139f7125d0e12dccb865869c33f888c432f809507ab34cc3e80d0f92ac"} Nov 25 06:05:53 crc kubenswrapper[4708]: I1125 06:05:53.922510 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerStarted","Data":"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa"} Nov 25 06:05:55 crc kubenswrapper[4708]: I1125 06:05:55.946533 4708 generic.go:334] "Generic (PLEG): container finished" podID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerID="c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa" exitCode=0 Nov 25 06:05:55 crc kubenswrapper[4708]: I1125 06:05:55.946570 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerDied","Data":"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa"} Nov 25 06:05:56 crc kubenswrapper[4708]: I1125 06:05:56.973016 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerStarted","Data":"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131"} Nov 25 06:05:56 crc kubenswrapper[4708]: I1125 06:05:56.992864 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-smlw2" podStartSLOduration=2.525400837 podStartE2EDuration="5.992843701s" podCreationTimestamp="2025-11-25 06:05:51 +0000 UTC" firstStartedPulling="2025-11-25 06:05:52.914327857 +0000 UTC m=+1494.323161243" lastFinishedPulling="2025-11-25 06:05:56.381770721 +0000 UTC m=+1497.790604107" observedRunningTime="2025-11-25 06:05:56.988062239 +0000 UTC m=+1498.396895624" watchObservedRunningTime="2025-11-25 06:05:56.992843701 +0000 UTC m=+1498.401677088" Nov 25 06:05:59 crc kubenswrapper[4708]: I1125 06:05:59.894920 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:05:59 crc kubenswrapper[4708]: E1125 06:05:59.895791 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:06:00 crc kubenswrapper[4708]: I1125 06:06:00.877265 4708 scope.go:117] "RemoveContainer" containerID="3977e9d7662fa5bf82ec5c52c5415bf8e0eb59e913711fe590c89f19eae60060" Nov 25 06:06:00 crc kubenswrapper[4708]: I1125 06:06:00.924007 4708 scope.go:117] "RemoveContainer" containerID="e7ff2d9e52ffd699b8820396f8c4afb7534c308c48c596fc2d0b5aa1b25b8abe" Nov 25 06:06:00 crc kubenswrapper[4708]: I1125 06:06:00.954059 4708 scope.go:117] "RemoveContainer" containerID="92d5c42cadaafa2a870118c54a28e34b934a292459ec7befc5120cf2cc463cf0" Nov 25 06:06:00 crc kubenswrapper[4708]: I1125 06:06:00.992636 4708 scope.go:117] "RemoveContainer" containerID="37ab3898ffc69fdbb2cd13dd6500e287093eea52fa83a90c64f0f8f0323b8664" Nov 25 06:06:01 crc kubenswrapper[4708]: I1125 06:06:01.039946 4708 scope.go:117] "RemoveContainer" containerID="e6e10051504f7f2c8358ecf0b9dc8076c4822fabbb4930beb6a6aacffc239809" Nov 25 06:06:01 crc kubenswrapper[4708]: I1125 06:06:01.794622 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:01 crc kubenswrapper[4708]: I1125 06:06:01.794689 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:01 crc kubenswrapper[4708]: I1125 06:06:01.834561 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:02 crc kubenswrapper[4708]: I1125 06:06:02.027116 4708 generic.go:334] "Generic (PLEG): container finished" podID="87296b4d-6e9f-4206-9d0f-78dd40e51e17" containerID="579560a0e9a1e30cf10588d2bada08120ec252aacb6534a285afea854cfb4b8b" exitCode=0 Nov 25 06:06:02 crc kubenswrapper[4708]: I1125 06:06:02.027214 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" event={"ID":"87296b4d-6e9f-4206-9d0f-78dd40e51e17","Type":"ContainerDied","Data":"579560a0e9a1e30cf10588d2bada08120ec252aacb6534a285afea854cfb4b8b"} Nov 25 06:06:02 crc kubenswrapper[4708]: I1125 06:06:02.069092 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:02 crc kubenswrapper[4708]: I1125 06:06:02.110096 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.389667 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.574588 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwx5d\" (UniqueName: \"kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d\") pod \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.574624 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory\") pod \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.574707 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key\") pod \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\" (UID: \"87296b4d-6e9f-4206-9d0f-78dd40e51e17\") " Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.579762 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d" (OuterVolumeSpecName: "kube-api-access-gwx5d") pod "87296b4d-6e9f-4206-9d0f-78dd40e51e17" (UID: "87296b4d-6e9f-4206-9d0f-78dd40e51e17"). InnerVolumeSpecName "kube-api-access-gwx5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.598962 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory" (OuterVolumeSpecName: "inventory") pod "87296b4d-6e9f-4206-9d0f-78dd40e51e17" (UID: "87296b4d-6e9f-4206-9d0f-78dd40e51e17"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.599071 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "87296b4d-6e9f-4206-9d0f-78dd40e51e17" (UID: "87296b4d-6e9f-4206-9d0f-78dd40e51e17"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.677580 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.677607 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwx5d\" (UniqueName: \"kubernetes.io/projected/87296b4d-6e9f-4206-9d0f-78dd40e51e17-kube-api-access-gwx5d\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:03 crc kubenswrapper[4708]: I1125 06:06:03.677622 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/87296b4d-6e9f-4206-9d0f-78dd40e51e17-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.047768 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" event={"ID":"87296b4d-6e9f-4206-9d0f-78dd40e51e17","Type":"ContainerDied","Data":"e490b4bf5caadeb34f9d51c400cdfe8a1b2a5ce7ad57c30f787b7fad83eb0ef9"} Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.047795 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.047822 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e490b4bf5caadeb34f9d51c400cdfe8a1b2a5ce7ad57c30f787b7fad83eb0ef9" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.047940 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-smlw2" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="registry-server" containerID="cri-o://9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131" gracePeriod=2 Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.118932 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g"] Nov 25 06:06:04 crc kubenswrapper[4708]: E1125 06:06:04.119434 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87296b4d-6e9f-4206-9d0f-78dd40e51e17" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.119451 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="87296b4d-6e9f-4206-9d0f-78dd40e51e17" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.119628 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="87296b4d-6e9f-4206-9d0f-78dd40e51e17" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.121200 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.133452 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.142658 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.142667 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.142732 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.157654 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g"] Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.188816 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.188897 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.188975 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4sbw\" (UniqueName: \"kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.292730 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.292804 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.292872 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4sbw\" (UniqueName: \"kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.301782 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.301900 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.309045 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4sbw\" (UniqueName: \"kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-4w55g\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.433406 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.452926 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.496194 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities\") pod \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.496302 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content\") pod \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.496462 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdfgt\" (UniqueName: \"kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt\") pod \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\" (UID: \"4745a0a8-7b6f-41fd-ae44-28fd233425a4\") " Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.498260 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities" (OuterVolumeSpecName: "utilities") pod "4745a0a8-7b6f-41fd-ae44-28fd233425a4" (UID: "4745a0a8-7b6f-41fd-ae44-28fd233425a4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.501265 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt" (OuterVolumeSpecName: "kube-api-access-fdfgt") pod "4745a0a8-7b6f-41fd-ae44-28fd233425a4" (UID: "4745a0a8-7b6f-41fd-ae44-28fd233425a4"). InnerVolumeSpecName "kube-api-access-fdfgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.579641 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4745a0a8-7b6f-41fd-ae44-28fd233425a4" (UID: "4745a0a8-7b6f-41fd-ae44-28fd233425a4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.598771 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.598795 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4745a0a8-7b6f-41fd-ae44-28fd233425a4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.598806 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdfgt\" (UniqueName: \"kubernetes.io/projected/4745a0a8-7b6f-41fd-ae44-28fd233425a4-kube-api-access-fdfgt\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:04 crc kubenswrapper[4708]: I1125 06:06:04.927140 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g"] Nov 25 06:06:04 crc kubenswrapper[4708]: W1125 06:06:04.931000 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec676a79_152d_4660_854e_5e447bd52ff2.slice/crio-20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512 WatchSource:0}: Error finding container 20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512: Status 404 returned error can't find the container with id 20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512 Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.057133 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" event={"ID":"ec676a79-152d-4660-854e-5e447bd52ff2","Type":"ContainerStarted","Data":"20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512"} Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.060600 4708 generic.go:334] "Generic (PLEG): container finished" podID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerID="9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131" exitCode=0 Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.060655 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerDied","Data":"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131"} Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.060671 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-smlw2" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.060692 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-smlw2" event={"ID":"4745a0a8-7b6f-41fd-ae44-28fd233425a4","Type":"ContainerDied","Data":"ecbde0139f7125d0e12dccb865869c33f888c432f809507ab34cc3e80d0f92ac"} Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.060712 4708 scope.go:117] "RemoveContainer" containerID="9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.083279 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.084812 4708 scope.go:117] "RemoveContainer" containerID="c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.089566 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-smlw2"] Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.103513 4708 scope.go:117] "RemoveContainer" containerID="ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.120718 4708 scope.go:117] "RemoveContainer" containerID="9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131" Nov 25 06:06:05 crc kubenswrapper[4708]: E1125 06:06:05.121166 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131\": container with ID starting with 9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131 not found: ID does not exist" containerID="9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.121238 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131"} err="failed to get container status \"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131\": rpc error: code = NotFound desc = could not find container \"9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131\": container with ID starting with 9a563e345a9091ad2e997d9f71f3b49e9adf899eb51dc953909369722b3bb131 not found: ID does not exist" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.121284 4708 scope.go:117] "RemoveContainer" containerID="c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa" Nov 25 06:06:05 crc kubenswrapper[4708]: E1125 06:06:05.121851 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa\": container with ID starting with c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa not found: ID does not exist" containerID="c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.121880 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa"} err="failed to get container status \"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa\": rpc error: code = NotFound desc = could not find container \"c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa\": container with ID starting with c59cdefb15d00ecd4973ef1f2946d2ce8bdb740d8888e6a2a61a6f2ad2f16faa not found: ID does not exist" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.121897 4708 scope.go:117] "RemoveContainer" containerID="ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97" Nov 25 06:06:05 crc kubenswrapper[4708]: E1125 06:06:05.122247 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97\": container with ID starting with ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97 not found: ID does not exist" containerID="ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97" Nov 25 06:06:05 crc kubenswrapper[4708]: I1125 06:06:05.122346 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97"} err="failed to get container status \"ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97\": rpc error: code = NotFound desc = could not find container \"ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97\": container with ID starting with ad6e54501dc370f54ac9a84fdb9a5dd8720986daec936a5a24a8b8e6380bab97 not found: ID does not exist" Nov 25 06:06:06 crc kubenswrapper[4708]: I1125 06:06:06.094544 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" event={"ID":"ec676a79-152d-4660-854e-5e447bd52ff2","Type":"ContainerStarted","Data":"31436e60afa57475b946a5ea996d3dbabec89e3dc932b285522d38509a468f67"} Nov 25 06:06:06 crc kubenswrapper[4708]: I1125 06:06:06.111133 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" podStartSLOduration=1.578904955 podStartE2EDuration="2.111114404s" podCreationTimestamp="2025-11-25 06:06:04 +0000 UTC" firstStartedPulling="2025-11-25 06:06:04.933510542 +0000 UTC m=+1506.342343928" lastFinishedPulling="2025-11-25 06:06:05.465719991 +0000 UTC m=+1506.874553377" observedRunningTime="2025-11-25 06:06:06.109492518 +0000 UTC m=+1507.518325904" watchObservedRunningTime="2025-11-25 06:06:06.111114404 +0000 UTC m=+1507.519947790" Nov 25 06:06:06 crc kubenswrapper[4708]: I1125 06:06:06.904515 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" path="/var/lib/kubelet/pods/4745a0a8-7b6f-41fd-ae44-28fd233425a4/volumes" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.481774 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:07 crc kubenswrapper[4708]: E1125 06:06:07.482626 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="extract-utilities" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.482641 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="extract-utilities" Nov 25 06:06:07 crc kubenswrapper[4708]: E1125 06:06:07.482653 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="registry-server" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.482659 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="registry-server" Nov 25 06:06:07 crc kubenswrapper[4708]: E1125 06:06:07.482676 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="extract-content" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.482683 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="extract-content" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.482879 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4745a0a8-7b6f-41fd-ae44-28fd233425a4" containerName="registry-server" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.484351 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.493992 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.655925 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46tpj\" (UniqueName: \"kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.656015 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.656286 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.758662 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.758907 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46tpj\" (UniqueName: \"kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.758985 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.759085 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.759388 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.779934 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46tpj\" (UniqueName: \"kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj\") pod \"community-operators-zlww4\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:07 crc kubenswrapper[4708]: I1125 06:06:07.799209 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:08 crc kubenswrapper[4708]: I1125 06:06:08.256812 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:08 crc kubenswrapper[4708]: W1125 06:06:08.259204 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2b719cc_8d1b_492e_ad2f_07b6faf3d73b.slice/crio-06380e57dcd6aae2c8c2a5eb8029ef7738b9cc30e3c0793eec60ea7ef645e5b8 WatchSource:0}: Error finding container 06380e57dcd6aae2c8c2a5eb8029ef7738b9cc30e3c0793eec60ea7ef645e5b8: Status 404 returned error can't find the container with id 06380e57dcd6aae2c8c2a5eb8029ef7738b9cc30e3c0793eec60ea7ef645e5b8 Nov 25 06:06:09 crc kubenswrapper[4708]: I1125 06:06:09.141619 4708 generic.go:334] "Generic (PLEG): container finished" podID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerID="5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239" exitCode=0 Nov 25 06:06:09 crc kubenswrapper[4708]: I1125 06:06:09.141754 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerDied","Data":"5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239"} Nov 25 06:06:09 crc kubenswrapper[4708]: I1125 06:06:09.142184 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerStarted","Data":"06380e57dcd6aae2c8c2a5eb8029ef7738b9cc30e3c0793eec60ea7ef645e5b8"} Nov 25 06:06:10 crc kubenswrapper[4708]: I1125 06:06:10.153007 4708 generic.go:334] "Generic (PLEG): container finished" podID="ec676a79-152d-4660-854e-5e447bd52ff2" containerID="31436e60afa57475b946a5ea996d3dbabec89e3dc932b285522d38509a468f67" exitCode=0 Nov 25 06:06:10 crc kubenswrapper[4708]: I1125 06:06:10.153093 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" event={"ID":"ec676a79-152d-4660-854e-5e447bd52ff2","Type":"ContainerDied","Data":"31436e60afa57475b946a5ea996d3dbabec89e3dc932b285522d38509a468f67"} Nov 25 06:06:10 crc kubenswrapper[4708]: I1125 06:06:10.158218 4708 generic.go:334] "Generic (PLEG): container finished" podID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerID="71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750" exitCode=0 Nov 25 06:06:10 crc kubenswrapper[4708]: I1125 06:06:10.158278 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerDied","Data":"71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750"} Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.173294 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerStarted","Data":"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e"} Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.214762 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zlww4" podStartSLOduration=2.720313838 podStartE2EDuration="4.214741156s" podCreationTimestamp="2025-11-25 06:06:07 +0000 UTC" firstStartedPulling="2025-11-25 06:06:09.144085607 +0000 UTC m=+1510.552918993" lastFinishedPulling="2025-11-25 06:06:10.638512925 +0000 UTC m=+1512.047346311" observedRunningTime="2025-11-25 06:06:11.204235784 +0000 UTC m=+1512.613069171" watchObservedRunningTime="2025-11-25 06:06:11.214741156 +0000 UTC m=+1512.623574542" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.477377 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.642361 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory\") pod \"ec676a79-152d-4660-854e-5e447bd52ff2\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.642538 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4sbw\" (UniqueName: \"kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw\") pod \"ec676a79-152d-4660-854e-5e447bd52ff2\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.642620 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key\") pod \"ec676a79-152d-4660-854e-5e447bd52ff2\" (UID: \"ec676a79-152d-4660-854e-5e447bd52ff2\") " Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.648683 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw" (OuterVolumeSpecName: "kube-api-access-l4sbw") pod "ec676a79-152d-4660-854e-5e447bd52ff2" (UID: "ec676a79-152d-4660-854e-5e447bd52ff2"). InnerVolumeSpecName "kube-api-access-l4sbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.669807 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory" (OuterVolumeSpecName: "inventory") pod "ec676a79-152d-4660-854e-5e447bd52ff2" (UID: "ec676a79-152d-4660-854e-5e447bd52ff2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.671900 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec676a79-152d-4660-854e-5e447bd52ff2" (UID: "ec676a79-152d-4660-854e-5e447bd52ff2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.745961 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.746003 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec676a79-152d-4660-854e-5e447bd52ff2-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:11 crc kubenswrapper[4708]: I1125 06:06:11.746017 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4sbw\" (UniqueName: \"kubernetes.io/projected/ec676a79-152d-4660-854e-5e447bd52ff2-kube-api-access-l4sbw\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.187077 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" event={"ID":"ec676a79-152d-4660-854e-5e447bd52ff2","Type":"ContainerDied","Data":"20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512"} Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.187602 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20cf433650cd1ec049b9e8731289f99106506c69bb64bb6600e426bd86c6d512" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.187106 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-4w55g" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.252133 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94"] Nov 25 06:06:12 crc kubenswrapper[4708]: E1125 06:06:12.252601 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec676a79-152d-4660-854e-5e447bd52ff2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.252625 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec676a79-152d-4660-854e-5e447bd52ff2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.252866 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec676a79-152d-4660-854e-5e447bd52ff2" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.253539 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.255624 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.255749 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.255892 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm7hs\" (UniqueName: \"kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.259109 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.259160 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.259479 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.259473 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.270690 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94"] Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.357585 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.357647 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.357713 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm7hs\" (UniqueName: \"kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.361176 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.362052 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.372293 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm7hs\" (UniqueName: \"kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-4lx94\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:12 crc kubenswrapper[4708]: I1125 06:06:12.572239 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:13 crc kubenswrapper[4708]: I1125 06:06:13.045980 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94"] Nov 25 06:06:13 crc kubenswrapper[4708]: W1125 06:06:13.048737 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded0b5403_c990_48b9_b3cc_8fe8455360a4.slice/crio-ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92 WatchSource:0}: Error finding container ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92: Status 404 returned error can't find the container with id ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92 Nov 25 06:06:13 crc kubenswrapper[4708]: I1125 06:06:13.197916 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" event={"ID":"ed0b5403-c990-48b9-b3cc-8fe8455360a4","Type":"ContainerStarted","Data":"ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92"} Nov 25 06:06:14 crc kubenswrapper[4708]: I1125 06:06:14.208549 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" event={"ID":"ed0b5403-c990-48b9-b3cc-8fe8455360a4","Type":"ContainerStarted","Data":"dede5bcfd15f7a4cc88bc301c468a3be338d520f024a4d139c53b492bd5cbdc6"} Nov 25 06:06:14 crc kubenswrapper[4708]: I1125 06:06:14.225152 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" podStartSLOduration=1.694797813 podStartE2EDuration="2.225121367s" podCreationTimestamp="2025-11-25 06:06:12 +0000 UTC" firstStartedPulling="2025-11-25 06:06:13.050914497 +0000 UTC m=+1514.459747883" lastFinishedPulling="2025-11-25 06:06:13.581238051 +0000 UTC m=+1514.990071437" observedRunningTime="2025-11-25 06:06:14.222582922 +0000 UTC m=+1515.631416319" watchObservedRunningTime="2025-11-25 06:06:14.225121367 +0000 UTC m=+1515.633954753" Nov 25 06:06:14 crc kubenswrapper[4708]: I1125 06:06:14.894143 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:06:14 crc kubenswrapper[4708]: E1125 06:06:14.894662 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:06:17 crc kubenswrapper[4708]: I1125 06:06:17.800276 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:17 crc kubenswrapper[4708]: I1125 06:06:17.800660 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:17 crc kubenswrapper[4708]: I1125 06:06:17.832631 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:18 crc kubenswrapper[4708]: I1125 06:06:18.284455 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:18 crc kubenswrapper[4708]: I1125 06:06:18.322503 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:20 crc kubenswrapper[4708]: I1125 06:06:20.038144 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0add-account-create-cxk7s"] Nov 25 06:06:20 crc kubenswrapper[4708]: I1125 06:06:20.046354 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0add-account-create-cxk7s"] Nov 25 06:06:20 crc kubenswrapper[4708]: I1125 06:06:20.260856 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zlww4" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="registry-server" containerID="cri-o://d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e" gracePeriod=2 Nov 25 06:06:20 crc kubenswrapper[4708]: I1125 06:06:20.905614 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="099277b8-d95e-41b6-b0d4-33f6346752c5" path="/var/lib/kubelet/pods/099277b8-d95e-41b6-b0d4-33f6346752c5/volumes" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.034202 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-5bf76"] Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.043730 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-084d-account-create-gx5q9"] Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.050287 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-084d-account-create-gx5q9"] Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.055391 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-5bf76"] Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.167028 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.269944 4708 generic.go:334] "Generic (PLEG): container finished" podID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerID="d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e" exitCode=0 Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.269984 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerDied","Data":"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e"} Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.270013 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlww4" event={"ID":"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b","Type":"ContainerDied","Data":"06380e57dcd6aae2c8c2a5eb8029ef7738b9cc30e3c0793eec60ea7ef645e5b8"} Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.270023 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlww4" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.270032 4708 scope.go:117] "RemoveContainer" containerID="d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.298015 4708 scope.go:117] "RemoveContainer" containerID="71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.317291 4708 scope.go:117] "RemoveContainer" containerID="5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.340226 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46tpj\" (UniqueName: \"kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj\") pod \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.340274 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content\") pod \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.340299 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities\") pod \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\" (UID: \"b2b719cc-8d1b-492e-ad2f-07b6faf3d73b\") " Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.341365 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities" (OuterVolumeSpecName: "utilities") pod "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" (UID: "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.341828 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.345345 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj" (OuterVolumeSpecName: "kube-api-access-46tpj") pod "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" (UID: "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b"). InnerVolumeSpecName "kube-api-access-46tpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.346038 4708 scope.go:117] "RemoveContainer" containerID="d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e" Nov 25 06:06:21 crc kubenswrapper[4708]: E1125 06:06:21.346515 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e\": container with ID starting with d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e not found: ID does not exist" containerID="d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.346563 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e"} err="failed to get container status \"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e\": rpc error: code = NotFound desc = could not find container \"d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e\": container with ID starting with d7f82d84b1b98badb0d10603124e04efc06c83021f7d22e5249d9a3ff282b11e not found: ID does not exist" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.346588 4708 scope.go:117] "RemoveContainer" containerID="71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750" Nov 25 06:06:21 crc kubenswrapper[4708]: E1125 06:06:21.346934 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750\": container with ID starting with 71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750 not found: ID does not exist" containerID="71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.346960 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750"} err="failed to get container status \"71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750\": rpc error: code = NotFound desc = could not find container \"71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750\": container with ID starting with 71457d696b9d6c5199478c064f37a05974dcf3539867b85ea4dbc91205270750 not found: ID does not exist" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.346982 4708 scope.go:117] "RemoveContainer" containerID="5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239" Nov 25 06:06:21 crc kubenswrapper[4708]: E1125 06:06:21.347269 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239\": container with ID starting with 5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239 not found: ID does not exist" containerID="5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.347508 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239"} err="failed to get container status \"5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239\": rpc error: code = NotFound desc = could not find container \"5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239\": container with ID starting with 5255694f4538994cee762fb207c71c13de7dd2e8f6bdb9f483c204ad133d0239 not found: ID does not exist" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.380637 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" (UID: "b2b719cc-8d1b-492e-ad2f-07b6faf3d73b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.444081 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46tpj\" (UniqueName: \"kubernetes.io/projected/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-kube-api-access-46tpj\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.444115 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.605326 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:21 crc kubenswrapper[4708]: I1125 06:06:21.611907 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zlww4"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.027336 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-53a6-account-create-fwcmk"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.033626 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jms52"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.042704 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-k52w2"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.049740 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-53a6-account-create-fwcmk"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.054578 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-k52w2"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.059173 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jms52"] Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.902950 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12614737-33a5-4a01-bd4a-91e9e41e0022" path="/var/lib/kubelet/pods/12614737-33a5-4a01-bd4a-91e9e41e0022/volumes" Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.903563 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b06819a-2b30-4c4a-aa31-97e1482aeab1" path="/var/lib/kubelet/pods/4b06819a-2b30-4c4a-aa31-97e1482aeab1/volumes" Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.904077 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99663ec8-6afb-455e-8fbb-ef9c015b6c60" path="/var/lib/kubelet/pods/99663ec8-6afb-455e-8fbb-ef9c015b6c60/volumes" Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.904621 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a91604ea-9b5a-4e39-96ec-abb8302360a0" path="/var/lib/kubelet/pods/a91604ea-9b5a-4e39-96ec-abb8302360a0/volumes" Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.905629 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9957cda-e4de-40f4-87fc-ce0dfd166505" path="/var/lib/kubelet/pods/a9957cda-e4de-40f4-87fc-ce0dfd166505/volumes" Nov 25 06:06:22 crc kubenswrapper[4708]: I1125 06:06:22.906118 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" path="/var/lib/kubelet/pods/b2b719cc-8d1b-492e-ad2f-07b6faf3d73b/volumes" Nov 25 06:06:26 crc kubenswrapper[4708]: I1125 06:06:26.894033 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:06:26 crc kubenswrapper[4708]: E1125 06:06:26.894697 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:06:38 crc kubenswrapper[4708]: I1125 06:06:38.899564 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:06:38 crc kubenswrapper[4708]: E1125 06:06:38.902555 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:06:40 crc kubenswrapper[4708]: I1125 06:06:40.443825 4708 generic.go:334] "Generic (PLEG): container finished" podID="ed0b5403-c990-48b9-b3cc-8fe8455360a4" containerID="dede5bcfd15f7a4cc88bc301c468a3be338d520f024a4d139c53b492bd5cbdc6" exitCode=0 Nov 25 06:06:40 crc kubenswrapper[4708]: I1125 06:06:40.443920 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" event={"ID":"ed0b5403-c990-48b9-b3cc-8fe8455360a4","Type":"ContainerDied","Data":"dede5bcfd15f7a4cc88bc301c468a3be338d520f024a4d139c53b492bd5cbdc6"} Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.825540 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.950394 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key\") pod \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.950544 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm7hs\" (UniqueName: \"kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs\") pod \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.950670 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory\") pod \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\" (UID: \"ed0b5403-c990-48b9-b3cc-8fe8455360a4\") " Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.955772 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs" (OuterVolumeSpecName: "kube-api-access-gm7hs") pod "ed0b5403-c990-48b9-b3cc-8fe8455360a4" (UID: "ed0b5403-c990-48b9-b3cc-8fe8455360a4"). InnerVolumeSpecName "kube-api-access-gm7hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.977126 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ed0b5403-c990-48b9-b3cc-8fe8455360a4" (UID: "ed0b5403-c990-48b9-b3cc-8fe8455360a4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:41 crc kubenswrapper[4708]: I1125 06:06:41.978764 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory" (OuterVolumeSpecName: "inventory") pod "ed0b5403-c990-48b9-b3cc-8fe8455360a4" (UID: "ed0b5403-c990-48b9-b3cc-8fe8455360a4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.055284 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.055317 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm7hs\" (UniqueName: \"kubernetes.io/projected/ed0b5403-c990-48b9-b3cc-8fe8455360a4-kube-api-access-gm7hs\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.055331 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed0b5403-c990-48b9-b3cc-8fe8455360a4-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.465858 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" event={"ID":"ed0b5403-c990-48b9-b3cc-8fe8455360a4","Type":"ContainerDied","Data":"ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92"} Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.465917 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba442d08b47ee0596cdf0e8f2ac09ec093e78ec645cf71da3624fe21cf5d3e92" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.465995 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-4lx94" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530356 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n"] Nov 25 06:06:42 crc kubenswrapper[4708]: E1125 06:06:42.530737 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed0b5403-c990-48b9-b3cc-8fe8455360a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530757 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed0b5403-c990-48b9-b3cc-8fe8455360a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:42 crc kubenswrapper[4708]: E1125 06:06:42.530771 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="registry-server" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530778 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="registry-server" Nov 25 06:06:42 crc kubenswrapper[4708]: E1125 06:06:42.530798 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="extract-utilities" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530804 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="extract-utilities" Nov 25 06:06:42 crc kubenswrapper[4708]: E1125 06:06:42.530816 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="extract-content" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530822 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="extract-content" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.530995 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed0b5403-c990-48b9-b3cc-8fe8455360a4" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.531011 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2b719cc-8d1b-492e-ad2f-07b6faf3d73b" containerName="registry-server" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.531652 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.533841 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.534498 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.534789 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.536143 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.546248 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n"] Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.565646 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.565698 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.565774 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwdzn\" (UniqueName: \"kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.667812 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.667872 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.667928 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwdzn\" (UniqueName: \"kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.672018 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.672417 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.680484 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwdzn\" (UniqueName: \"kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:42 crc kubenswrapper[4708]: I1125 06:06:42.846703 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:06:43 crc kubenswrapper[4708]: I1125 06:06:43.327945 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n"] Nov 25 06:06:43 crc kubenswrapper[4708]: I1125 06:06:43.477816 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" event={"ID":"9084cfbd-0f77-4a71-b265-fdbe8ac68bef","Type":"ContainerStarted","Data":"aedb1697393d014ce056ebebe09496dbaa6acdbe0a44f1f2f2fc31f0b580bc65"} Nov 25 06:06:44 crc kubenswrapper[4708]: I1125 06:06:44.486985 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" event={"ID":"9084cfbd-0f77-4a71-b265-fdbe8ac68bef","Type":"ContainerStarted","Data":"205db9991cf666680f2215c2657de5954c9404da83abb3cb2ebc1a270f6c0404"} Nov 25 06:06:44 crc kubenswrapper[4708]: I1125 06:06:44.503961 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" podStartSLOduration=2.012698321 podStartE2EDuration="2.50393953s" podCreationTimestamp="2025-11-25 06:06:42 +0000 UTC" firstStartedPulling="2025-11-25 06:06:43.331894805 +0000 UTC m=+1544.740728191" lastFinishedPulling="2025-11-25 06:06:43.823136024 +0000 UTC m=+1545.231969400" observedRunningTime="2025-11-25 06:06:44.501218501 +0000 UTC m=+1545.910051887" watchObservedRunningTime="2025-11-25 06:06:44.50393953 +0000 UTC m=+1545.912772916" Nov 25 06:06:46 crc kubenswrapper[4708]: I1125 06:06:46.036828 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vhfrh"] Nov 25 06:06:46 crc kubenswrapper[4708]: I1125 06:06:46.042175 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-vhfrh"] Nov 25 06:06:46 crc kubenswrapper[4708]: I1125 06:06:46.905234 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ba92183-cc87-4b0f-8b71-8be8d39be23c" path="/var/lib/kubelet/pods/7ba92183-cc87-4b0f-8b71-8be8d39be23c/volumes" Nov 25 06:06:49 crc kubenswrapper[4708]: I1125 06:06:49.893921 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:06:49 crc kubenswrapper[4708]: E1125 06:06:49.894663 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.178791 4708 scope.go:117] "RemoveContainer" containerID="9b99ce75bdfdeb5995cbb6a828c83d5dadb2dfad727065db138261823ea11f50" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.209855 4708 scope.go:117] "RemoveContainer" containerID="4969ab9724ca3c3f14130a3efdd9fa73c081d7e796b5efabf94f5301c2c5259e" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.243564 4708 scope.go:117] "RemoveContainer" containerID="36369ecde8c56951a3357e3f969cdfacab7e267a99860b9d9ef8e816a9b59362" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.274653 4708 scope.go:117] "RemoveContainer" containerID="237199b481c43a69bbff1a05921d2624cc211bf25c7f1138db9af1168796e82a" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.303230 4708 scope.go:117] "RemoveContainer" containerID="946fcc0147e7f344e51461ce86a2b8012291b42ea3d6b1254bfa7b921e349f55" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.360534 4708 scope.go:117] "RemoveContainer" containerID="bdd00b3b0105ac4475633b45b1144b71cd968505cdb599227905d7c3a349edc1" Nov 25 06:07:01 crc kubenswrapper[4708]: I1125 06:07:01.394913 4708 scope.go:117] "RemoveContainer" containerID="e29f0bc44d62fee9a60e174d60c2dac676316f4cd8170eb024f4df3196ad31e8" Nov 25 06:07:02 crc kubenswrapper[4708]: I1125 06:07:02.025878 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbbt5"] Nov 25 06:07:02 crc kubenswrapper[4708]: I1125 06:07:02.031211 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vbbt5"] Nov 25 06:07:02 crc kubenswrapper[4708]: I1125 06:07:02.929695 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2654acf-c93c-4c8d-bd73-537b9c828615" path="/var/lib/kubelet/pods/f2654acf-c93c-4c8d-bd73-537b9c828615/volumes" Nov 25 06:07:03 crc kubenswrapper[4708]: I1125 06:07:03.024698 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-5gq69"] Nov 25 06:07:03 crc kubenswrapper[4708]: I1125 06:07:03.033202 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-5gq69"] Nov 25 06:07:04 crc kubenswrapper[4708]: I1125 06:07:04.893148 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:07:04 crc kubenswrapper[4708]: E1125 06:07:04.893785 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:07:04 crc kubenswrapper[4708]: I1125 06:07:04.904169 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="527c5c0e-4e68-4533-b9b0-e65cfd465ce7" path="/var/lib/kubelet/pods/527c5c0e-4e68-4533-b9b0-e65cfd465ce7/volumes" Nov 25 06:07:16 crc kubenswrapper[4708]: I1125 06:07:16.893368 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:07:16 crc kubenswrapper[4708]: E1125 06:07:16.894362 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:07:19 crc kubenswrapper[4708]: I1125 06:07:19.817363 4708 generic.go:334] "Generic (PLEG): container finished" podID="9084cfbd-0f77-4a71-b265-fdbe8ac68bef" containerID="205db9991cf666680f2215c2657de5954c9404da83abb3cb2ebc1a270f6c0404" exitCode=0 Nov 25 06:07:19 crc kubenswrapper[4708]: I1125 06:07:19.817465 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" event={"ID":"9084cfbd-0f77-4a71-b265-fdbe8ac68bef","Type":"ContainerDied","Data":"205db9991cf666680f2215c2657de5954c9404da83abb3cb2ebc1a270f6c0404"} Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.166015 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.322208 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key\") pod \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.322368 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory\") pod \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.322469 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qwdzn\" (UniqueName: \"kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn\") pod \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\" (UID: \"9084cfbd-0f77-4a71-b265-fdbe8ac68bef\") " Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.328780 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn" (OuterVolumeSpecName: "kube-api-access-qwdzn") pod "9084cfbd-0f77-4a71-b265-fdbe8ac68bef" (UID: "9084cfbd-0f77-4a71-b265-fdbe8ac68bef"). InnerVolumeSpecName "kube-api-access-qwdzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.352900 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory" (OuterVolumeSpecName: "inventory") pod "9084cfbd-0f77-4a71-b265-fdbe8ac68bef" (UID: "9084cfbd-0f77-4a71-b265-fdbe8ac68bef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.353342 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9084cfbd-0f77-4a71-b265-fdbe8ac68bef" (UID: "9084cfbd-0f77-4a71-b265-fdbe8ac68bef"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.424663 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.424723 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qwdzn\" (UniqueName: \"kubernetes.io/projected/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-kube-api-access-qwdzn\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.424740 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9084cfbd-0f77-4a71-b265-fdbe8ac68bef-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.864847 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" event={"ID":"9084cfbd-0f77-4a71-b265-fdbe8ac68bef","Type":"ContainerDied","Data":"aedb1697393d014ce056ebebe09496dbaa6acdbe0a44f1f2f2fc31f0b580bc65"} Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.865454 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aedb1697393d014ce056ebebe09496dbaa6acdbe0a44f1f2f2fc31f0b580bc65" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.865630 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.947637 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7t6t"] Nov 25 06:07:21 crc kubenswrapper[4708]: E1125 06:07:21.948104 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9084cfbd-0f77-4a71-b265-fdbe8ac68bef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.948124 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="9084cfbd-0f77-4a71-b265-fdbe8ac68bef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.948323 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="9084cfbd-0f77-4a71-b265-fdbe8ac68bef" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.960142 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.963499 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7t6t"] Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.963626 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.963779 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.963916 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:07:21 crc kubenswrapper[4708]: I1125 06:07:21.968100 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:07:22 crc kubenswrapper[4708]: E1125 06:07:22.011888 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9084cfbd_0f77_4a71_b265_fdbe8ac68bef.slice/crio-aedb1697393d014ce056ebebe09496dbaa6acdbe0a44f1f2f2fc31f0b580bc65\": RecentStats: unable to find data in memory cache]" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.143336 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.143781 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.143963 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn7wz\" (UniqueName: \"kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.245243 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.245306 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.245375 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn7wz\" (UniqueName: \"kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.251727 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.251839 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.263120 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn7wz\" (UniqueName: \"kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz\") pod \"ssh-known-hosts-edpm-deployment-t7t6t\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.284348 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.736249 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7t6t"] Nov 25 06:07:22 crc kubenswrapper[4708]: I1125 06:07:22.876700 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" event={"ID":"a2cb620c-a539-4928-9a66-6c08cf197798","Type":"ContainerStarted","Data":"3145d188aa015ace4ace02a3464835f376ea506f764e385798a5d3b00e722439"} Nov 25 06:07:23 crc kubenswrapper[4708]: I1125 06:07:23.888108 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" event={"ID":"a2cb620c-a539-4928-9a66-6c08cf197798","Type":"ContainerStarted","Data":"6bdd1b1aafa3c44b49eb4b7df2df898e6b090148970df1422b6b732d09c987c4"} Nov 25 06:07:23 crc kubenswrapper[4708]: I1125 06:07:23.904036 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" podStartSLOduration=2.363510521 podStartE2EDuration="2.904000678s" podCreationTimestamp="2025-11-25 06:07:21 +0000 UTC" firstStartedPulling="2025-11-25 06:07:22.740658468 +0000 UTC m=+1584.149491854" lastFinishedPulling="2025-11-25 06:07:23.281148624 +0000 UTC m=+1584.689982011" observedRunningTime="2025-11-25 06:07:23.900939829 +0000 UTC m=+1585.309773215" watchObservedRunningTime="2025-11-25 06:07:23.904000678 +0000 UTC m=+1585.312834064" Nov 25 06:07:28 crc kubenswrapper[4708]: I1125 06:07:28.942813 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2cb620c-a539-4928-9a66-6c08cf197798" containerID="6bdd1b1aafa3c44b49eb4b7df2df898e6b090148970df1422b6b732d09c987c4" exitCode=0 Nov 25 06:07:28 crc kubenswrapper[4708]: I1125 06:07:28.942907 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" event={"ID":"a2cb620c-a539-4928-9a66-6c08cf197798","Type":"ContainerDied","Data":"6bdd1b1aafa3c44b49eb4b7df2df898e6b090148970df1422b6b732d09c987c4"} Nov 25 06:07:29 crc kubenswrapper[4708]: I1125 06:07:29.899359 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:07:29 crc kubenswrapper[4708]: E1125 06:07:29.900759 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.295963 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.424468 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0\") pod \"a2cb620c-a539-4928-9a66-6c08cf197798\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.424581 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kn7wz\" (UniqueName: \"kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz\") pod \"a2cb620c-a539-4928-9a66-6c08cf197798\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.424693 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam\") pod \"a2cb620c-a539-4928-9a66-6c08cf197798\" (UID: \"a2cb620c-a539-4928-9a66-6c08cf197798\") " Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.456644 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz" (OuterVolumeSpecName: "kube-api-access-kn7wz") pod "a2cb620c-a539-4928-9a66-6c08cf197798" (UID: "a2cb620c-a539-4928-9a66-6c08cf197798"). InnerVolumeSpecName "kube-api-access-kn7wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.476576 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "a2cb620c-a539-4928-9a66-6c08cf197798" (UID: "a2cb620c-a539-4928-9a66-6c08cf197798"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.478843 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a2cb620c-a539-4928-9a66-6c08cf197798" (UID: "a2cb620c-a539-4928-9a66-6c08cf197798"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.527420 4708 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.527464 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kn7wz\" (UniqueName: \"kubernetes.io/projected/a2cb620c-a539-4928-9a66-6c08cf197798-kube-api-access-kn7wz\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.527480 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a2cb620c-a539-4928-9a66-6c08cf197798-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.962699 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" event={"ID":"a2cb620c-a539-4928-9a66-6c08cf197798","Type":"ContainerDied","Data":"3145d188aa015ace4ace02a3464835f376ea506f764e385798a5d3b00e722439"} Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.962768 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7t6t" Nov 25 06:07:30 crc kubenswrapper[4708]: I1125 06:07:30.962807 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3145d188aa015ace4ace02a3464835f376ea506f764e385798a5d3b00e722439" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.018553 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh"] Nov 25 06:07:31 crc kubenswrapper[4708]: E1125 06:07:31.018994 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2cb620c-a539-4928-9a66-6c08cf197798" containerName="ssh-known-hosts-edpm-deployment" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.019013 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2cb620c-a539-4928-9a66-6c08cf197798" containerName="ssh-known-hosts-edpm-deployment" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.019168 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2cb620c-a539-4928-9a66-6c08cf197798" containerName="ssh-known-hosts-edpm-deployment" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.019831 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.024255 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.025023 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.025202 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.026638 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.027974 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh"] Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.140469 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.140847 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tn8h\" (UniqueName: \"kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.141114 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.243944 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.244089 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tn8h\" (UniqueName: \"kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.244171 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.249609 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.249714 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.261166 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tn8h\" (UniqueName: \"kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-6k7zh\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.333802 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.816995 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh"] Nov 25 06:07:31 crc kubenswrapper[4708]: I1125 06:07:31.974115 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" event={"ID":"1ef0663b-adb0-4192-a17a-1643899f466f","Type":"ContainerStarted","Data":"fe3991006cf529ae4b08a2c1b86d541c40b149dc0c98acd1fc58b9668a95b944"} Nov 25 06:07:32 crc kubenswrapper[4708]: I1125 06:07:32.983431 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" event={"ID":"1ef0663b-adb0-4192-a17a-1643899f466f","Type":"ContainerStarted","Data":"fdf47badf8a8654085571c2145919d4a8246b04276f008cd8c03cd98cd59b52e"} Nov 25 06:07:32 crc kubenswrapper[4708]: I1125 06:07:32.996758 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" podStartSLOduration=1.460216926 podStartE2EDuration="1.99672602s" podCreationTimestamp="2025-11-25 06:07:31 +0000 UTC" firstStartedPulling="2025-11-25 06:07:31.821919169 +0000 UTC m=+1593.230752556" lastFinishedPulling="2025-11-25 06:07:32.358428274 +0000 UTC m=+1593.767261650" observedRunningTime="2025-11-25 06:07:32.996165834 +0000 UTC m=+1594.404999220" watchObservedRunningTime="2025-11-25 06:07:32.99672602 +0000 UTC m=+1594.405559407" Nov 25 06:07:39 crc kubenswrapper[4708]: I1125 06:07:39.035686 4708 generic.go:334] "Generic (PLEG): container finished" podID="1ef0663b-adb0-4192-a17a-1643899f466f" containerID="fdf47badf8a8654085571c2145919d4a8246b04276f008cd8c03cd98cd59b52e" exitCode=0 Nov 25 06:07:39 crc kubenswrapper[4708]: I1125 06:07:39.035736 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" event={"ID":"1ef0663b-adb0-4192-a17a-1643899f466f","Type":"ContainerDied","Data":"fdf47badf8a8654085571c2145919d4a8246b04276f008cd8c03cd98cd59b52e"} Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.371374 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.529818 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory\") pod \"1ef0663b-adb0-4192-a17a-1643899f466f\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.530200 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key\") pod \"1ef0663b-adb0-4192-a17a-1643899f466f\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.530631 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tn8h\" (UniqueName: \"kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h\") pod \"1ef0663b-adb0-4192-a17a-1643899f466f\" (UID: \"1ef0663b-adb0-4192-a17a-1643899f466f\") " Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.536892 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h" (OuterVolumeSpecName: "kube-api-access-2tn8h") pod "1ef0663b-adb0-4192-a17a-1643899f466f" (UID: "1ef0663b-adb0-4192-a17a-1643899f466f"). InnerVolumeSpecName "kube-api-access-2tn8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.555161 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory" (OuterVolumeSpecName: "inventory") pod "1ef0663b-adb0-4192-a17a-1643899f466f" (UID: "1ef0663b-adb0-4192-a17a-1643899f466f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.556180 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ef0663b-adb0-4192-a17a-1643899f466f" (UID: "1ef0663b-adb0-4192-a17a-1643899f466f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.633327 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tn8h\" (UniqueName: \"kubernetes.io/projected/1ef0663b-adb0-4192-a17a-1643899f466f-kube-api-access-2tn8h\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.633358 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:40 crc kubenswrapper[4708]: I1125 06:07:40.633370 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ef0663b-adb0-4192-a17a-1643899f466f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.058393 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" event={"ID":"1ef0663b-adb0-4192-a17a-1643899f466f","Type":"ContainerDied","Data":"fe3991006cf529ae4b08a2c1b86d541c40b149dc0c98acd1fc58b9668a95b944"} Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.058700 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe3991006cf529ae4b08a2c1b86d541c40b149dc0c98acd1fc58b9668a95b944" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.058471 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-6k7zh" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.121834 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6"] Nov 25 06:07:41 crc kubenswrapper[4708]: E1125 06:07:41.122549 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ef0663b-adb0-4192-a17a-1643899f466f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.122669 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ef0663b-adb0-4192-a17a-1643899f466f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.122991 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ef0663b-adb0-4192-a17a-1643899f466f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.123788 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.125692 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.125875 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.126219 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.126338 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.138991 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6"] Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.242819 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.242944 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crp2q\" (UniqueName: \"kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.243003 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.345123 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crp2q\" (UniqueName: \"kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.345214 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.345309 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.351495 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.351722 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.360717 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crp2q\" (UniqueName: \"kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.437560 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:41 crc kubenswrapper[4708]: I1125 06:07:41.908423 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6"] Nov 25 06:07:42 crc kubenswrapper[4708]: I1125 06:07:42.070021 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" event={"ID":"1ffad4ac-de8a-4f8c-8409-20b8553a4259","Type":"ContainerStarted","Data":"e5e8bc2a3552a6803e980f338208f40583a06b2a3dd54bb546fb5a0ec8d139e1"} Nov 25 06:07:43 crc kubenswrapper[4708]: I1125 06:07:43.087765 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" event={"ID":"1ffad4ac-de8a-4f8c-8409-20b8553a4259","Type":"ContainerStarted","Data":"81153e37734e6ebf4b5ed172eb8e0be57637170d538aac936ef92192f57eb5e3"} Nov 25 06:07:43 crc kubenswrapper[4708]: I1125 06:07:43.118439 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" podStartSLOduration=1.627025977 podStartE2EDuration="2.118419382s" podCreationTimestamp="2025-11-25 06:07:41 +0000 UTC" firstStartedPulling="2025-11-25 06:07:41.911793314 +0000 UTC m=+1603.320626700" lastFinishedPulling="2025-11-25 06:07:42.403186718 +0000 UTC m=+1603.812020105" observedRunningTime="2025-11-25 06:07:43.112900618 +0000 UTC m=+1604.521734005" watchObservedRunningTime="2025-11-25 06:07:43.118419382 +0000 UTC m=+1604.527252768" Nov 25 06:07:44 crc kubenswrapper[4708]: I1125 06:07:44.893755 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:07:44 crc kubenswrapper[4708]: E1125 06:07:44.895733 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:07:47 crc kubenswrapper[4708]: I1125 06:07:47.040300 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-d96zq"] Nov 25 06:07:47 crc kubenswrapper[4708]: I1125 06:07:47.047713 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-d96zq"] Nov 25 06:07:48 crc kubenswrapper[4708]: I1125 06:07:48.904031 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee0a911-c6e0-43b0-9b11-81f4d283dea4" path="/var/lib/kubelet/pods/cee0a911-c6e0-43b0-9b11-81f4d283dea4/volumes" Nov 25 06:07:50 crc kubenswrapper[4708]: I1125 06:07:50.155071 4708 generic.go:334] "Generic (PLEG): container finished" podID="1ffad4ac-de8a-4f8c-8409-20b8553a4259" containerID="81153e37734e6ebf4b5ed172eb8e0be57637170d538aac936ef92192f57eb5e3" exitCode=0 Nov 25 06:07:50 crc kubenswrapper[4708]: I1125 06:07:50.155131 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" event={"ID":"1ffad4ac-de8a-4f8c-8409-20b8553a4259","Type":"ContainerDied","Data":"81153e37734e6ebf4b5ed172eb8e0be57637170d538aac936ef92192f57eb5e3"} Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.520772 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.561904 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crp2q\" (UniqueName: \"kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q\") pod \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.562106 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key\") pod \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.562290 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory\") pod \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\" (UID: \"1ffad4ac-de8a-4f8c-8409-20b8553a4259\") " Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.567914 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q" (OuterVolumeSpecName: "kube-api-access-crp2q") pod "1ffad4ac-de8a-4f8c-8409-20b8553a4259" (UID: "1ffad4ac-de8a-4f8c-8409-20b8553a4259"). InnerVolumeSpecName "kube-api-access-crp2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.585956 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory" (OuterVolumeSpecName: "inventory") pod "1ffad4ac-de8a-4f8c-8409-20b8553a4259" (UID: "1ffad4ac-de8a-4f8c-8409-20b8553a4259"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.586960 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1ffad4ac-de8a-4f8c-8409-20b8553a4259" (UID: "1ffad4ac-de8a-4f8c-8409-20b8553a4259"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.664379 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.664413 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1ffad4ac-de8a-4f8c-8409-20b8553a4259-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:51 crc kubenswrapper[4708]: I1125 06:07:51.664426 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crp2q\" (UniqueName: \"kubernetes.io/projected/1ffad4ac-de8a-4f8c-8409-20b8553a4259-kube-api-access-crp2q\") on node \"crc\" DevicePath \"\"" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.174716 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" event={"ID":"1ffad4ac-de8a-4f8c-8409-20b8553a4259","Type":"ContainerDied","Data":"e5e8bc2a3552a6803e980f338208f40583a06b2a3dd54bb546fb5a0ec8d139e1"} Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.174763 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.174771 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5e8bc2a3552a6803e980f338208f40583a06b2a3dd54bb546fb5a0ec8d139e1" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.315751 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm"] Nov 25 06:07:52 crc kubenswrapper[4708]: E1125 06:07:52.316235 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ffad4ac-de8a-4f8c-8409-20b8553a4259" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.316257 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ffad4ac-de8a-4f8c-8409-20b8553a4259" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.316436 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ffad4ac-de8a-4f8c-8409-20b8553a4259" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.317148 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319471 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319639 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319662 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319643 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319760 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.319869 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.320723 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.320883 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.321849 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm"] Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.482919 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.482976 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483026 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483095 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483162 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483191 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483262 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483289 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483355 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483462 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483512 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483600 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483697 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64mpj\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.483745 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.585620 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64mpj\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586016 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586072 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586098 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586123 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586149 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586184 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586217 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586251 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586275 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586297 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586323 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586356 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.586393 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.594095 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.594457 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.594827 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.594948 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.594998 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.595466 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.595809 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.597244 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.598427 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.599085 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.599151 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.599165 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.599804 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.601282 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64mpj\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-244cm\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:52 crc kubenswrapper[4708]: I1125 06:07:52.637567 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:07:53 crc kubenswrapper[4708]: I1125 06:07:53.111832 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm"] Nov 25 06:07:53 crc kubenswrapper[4708]: I1125 06:07:53.181964 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" event={"ID":"a0c3d4bb-310d-4de3-84d8-67f8654706b7","Type":"ContainerStarted","Data":"62a2a582f0d3d54b6c5072dd4540abbd36784187e3eb92ceca5ab411acb98f42"} Nov 25 06:07:54 crc kubenswrapper[4708]: I1125 06:07:54.194077 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" event={"ID":"a0c3d4bb-310d-4de3-84d8-67f8654706b7","Type":"ContainerStarted","Data":"ddc405da9e9d93733b0c4402c115401f1622887600c32af3b9e501d20186233d"} Nov 25 06:07:54 crc kubenswrapper[4708]: I1125 06:07:54.221320 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" podStartSLOduration=1.602924239 podStartE2EDuration="2.221303221s" podCreationTimestamp="2025-11-25 06:07:52 +0000 UTC" firstStartedPulling="2025-11-25 06:07:53.115432509 +0000 UTC m=+1614.524265895" lastFinishedPulling="2025-11-25 06:07:53.733811491 +0000 UTC m=+1615.142644877" observedRunningTime="2025-11-25 06:07:54.215334 +0000 UTC m=+1615.624167386" watchObservedRunningTime="2025-11-25 06:07:54.221303221 +0000 UTC m=+1615.630136608" Nov 25 06:07:55 crc kubenswrapper[4708]: I1125 06:07:55.893052 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:07:55 crc kubenswrapper[4708]: E1125 06:07:55.893756 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:08:01 crc kubenswrapper[4708]: I1125 06:08:01.539279 4708 scope.go:117] "RemoveContainer" containerID="961cb4a067c64ddf40f3839f15f99e2ef29c9d639cb758cc8daf5b4deb9bef7b" Nov 25 06:08:01 crc kubenswrapper[4708]: I1125 06:08:01.587085 4708 scope.go:117] "RemoveContainer" containerID="b7130d42a3fdc3f2702b6ea2382f13f367557e7becf529cc8e8a14f39882f4a4" Nov 25 06:08:01 crc kubenswrapper[4708]: I1125 06:08:01.634119 4708 scope.go:117] "RemoveContainer" containerID="9ae42d476d5f2424caf80504e7ddcc8023c2bc5417abd3fb858475768879d891" Nov 25 06:08:08 crc kubenswrapper[4708]: I1125 06:08:08.898979 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:08:08 crc kubenswrapper[4708]: E1125 06:08:08.899882 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:08:19 crc kubenswrapper[4708]: I1125 06:08:19.893570 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:08:19 crc kubenswrapper[4708]: E1125 06:08:19.895452 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:08:20 crc kubenswrapper[4708]: I1125 06:08:20.444280 4708 generic.go:334] "Generic (PLEG): container finished" podID="a0c3d4bb-310d-4de3-84d8-67f8654706b7" containerID="ddc405da9e9d93733b0c4402c115401f1622887600c32af3b9e501d20186233d" exitCode=0 Nov 25 06:08:20 crc kubenswrapper[4708]: I1125 06:08:20.444390 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" event={"ID":"a0c3d4bb-310d-4de3-84d8-67f8654706b7","Type":"ContainerDied","Data":"ddc405da9e9d93733b0c4402c115401f1622887600c32af3b9e501d20186233d"} Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.796656 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.973403 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.973627 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.973658 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974137 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974205 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974606 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974699 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974755 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974836 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974901 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.974969 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.975009 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.975077 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64mpj\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.975136 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle\") pod \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\" (UID: \"a0c3d4bb-310d-4de3-84d8-67f8654706b7\") " Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.980443 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.981071 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.981351 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.981796 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.983504 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.983656 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.983912 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.984078 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.985168 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.985260 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.985777 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:21 crc kubenswrapper[4708]: I1125 06:08:21.986080 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj" (OuterVolumeSpecName: "kube-api-access-64mpj") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "kube-api-access-64mpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.004147 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.004638 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory" (OuterVolumeSpecName: "inventory") pod "a0c3d4bb-310d-4de3-84d8-67f8654706b7" (UID: "a0c3d4bb-310d-4de3-84d8-67f8654706b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078467 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078592 4708 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078666 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078722 4708 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078771 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078823 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078886 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078943 4708 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.078999 4708 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.079053 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.079106 4708 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.079161 4708 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.079223 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64mpj\" (UniqueName: \"kubernetes.io/projected/a0c3d4bb-310d-4de3-84d8-67f8654706b7-kube-api-access-64mpj\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.079280 4708 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a0c3d4bb-310d-4de3-84d8-67f8654706b7-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.467013 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" event={"ID":"a0c3d4bb-310d-4de3-84d8-67f8654706b7","Type":"ContainerDied","Data":"62a2a582f0d3d54b6c5072dd4540abbd36784187e3eb92ceca5ab411acb98f42"} Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.467072 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62a2a582f0d3d54b6c5072dd4540abbd36784187e3eb92ceca5ab411acb98f42" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.467423 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-244cm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.545706 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm"] Nov 25 06:08:22 crc kubenswrapper[4708]: E1125 06:08:22.546106 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0c3d4bb-310d-4de3-84d8-67f8654706b7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.546128 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0c3d4bb-310d-4de3-84d8-67f8654706b7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.546352 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0c3d4bb-310d-4de3-84d8-67f8654706b7" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.547031 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.548916 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.549253 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.549258 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.549272 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.549314 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.556062 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm"] Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.587642 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.587692 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsdrn\" (UniqueName: \"kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.587788 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.587831 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.587867 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.690648 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.691293 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.691340 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.691498 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.691555 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsdrn\" (UniqueName: \"kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.691877 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.696152 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.696582 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.697874 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.709419 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsdrn\" (UniqueName: \"kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kh7hm\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:22 crc kubenswrapper[4708]: I1125 06:08:22.861457 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:08:23 crc kubenswrapper[4708]: I1125 06:08:23.335673 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm"] Nov 25 06:08:23 crc kubenswrapper[4708]: I1125 06:08:23.476895 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" event={"ID":"4b251231-3104-491a-9dc1-bac1869d716d","Type":"ContainerStarted","Data":"58d01c124f215a3d213dd3bfc7ba3f76b496047a1ba7883312c730abd7b5b024"} Nov 25 06:08:24 crc kubenswrapper[4708]: I1125 06:08:24.487914 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" event={"ID":"4b251231-3104-491a-9dc1-bac1869d716d","Type":"ContainerStarted","Data":"31e4e4002eed47989787574dd210d76fb9bbd23fc91e851b81b186517a199702"} Nov 25 06:08:24 crc kubenswrapper[4708]: I1125 06:08:24.514119 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" podStartSLOduration=1.947496134 podStartE2EDuration="2.514096573s" podCreationTimestamp="2025-11-25 06:08:22 +0000 UTC" firstStartedPulling="2025-11-25 06:08:23.334060393 +0000 UTC m=+1644.742893780" lastFinishedPulling="2025-11-25 06:08:23.900660833 +0000 UTC m=+1645.309494219" observedRunningTime="2025-11-25 06:08:24.501493587 +0000 UTC m=+1645.910326974" watchObservedRunningTime="2025-11-25 06:08:24.514096573 +0000 UTC m=+1645.922929959" Nov 25 06:08:31 crc kubenswrapper[4708]: I1125 06:08:31.893782 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:08:31 crc kubenswrapper[4708]: E1125 06:08:31.894913 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:08:43 crc kubenswrapper[4708]: I1125 06:08:43.893936 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:08:43 crc kubenswrapper[4708]: E1125 06:08:43.895009 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:08:55 crc kubenswrapper[4708]: I1125 06:08:55.894081 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:08:55 crc kubenswrapper[4708]: E1125 06:08:55.895075 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:09:08 crc kubenswrapper[4708]: I1125 06:09:08.887814 4708 generic.go:334] "Generic (PLEG): container finished" podID="4b251231-3104-491a-9dc1-bac1869d716d" containerID="31e4e4002eed47989787574dd210d76fb9bbd23fc91e851b81b186517a199702" exitCode=0 Nov 25 06:09:08 crc kubenswrapper[4708]: I1125 06:09:08.887915 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" event={"ID":"4b251231-3104-491a-9dc1-bac1869d716d","Type":"ContainerDied","Data":"31e4e4002eed47989787574dd210d76fb9bbd23fc91e851b81b186517a199702"} Nov 25 06:09:08 crc kubenswrapper[4708]: I1125 06:09:08.893961 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:09:08 crc kubenswrapper[4708]: E1125 06:09:08.894280 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.191242 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.388722 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle\") pod \"4b251231-3104-491a-9dc1-bac1869d716d\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.389000 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0\") pod \"4b251231-3104-491a-9dc1-bac1869d716d\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.389034 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hsdrn\" (UniqueName: \"kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn\") pod \"4b251231-3104-491a-9dc1-bac1869d716d\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.389235 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key\") pod \"4b251231-3104-491a-9dc1-bac1869d716d\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.389355 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory\") pod \"4b251231-3104-491a-9dc1-bac1869d716d\" (UID: \"4b251231-3104-491a-9dc1-bac1869d716d\") " Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.395470 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4b251231-3104-491a-9dc1-bac1869d716d" (UID: "4b251231-3104-491a-9dc1-bac1869d716d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.395722 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn" (OuterVolumeSpecName: "kube-api-access-hsdrn") pod "4b251231-3104-491a-9dc1-bac1869d716d" (UID: "4b251231-3104-491a-9dc1-bac1869d716d"). InnerVolumeSpecName "kube-api-access-hsdrn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.412886 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory" (OuterVolumeSpecName: "inventory") pod "4b251231-3104-491a-9dc1-bac1869d716d" (UID: "4b251231-3104-491a-9dc1-bac1869d716d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.416830 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4b251231-3104-491a-9dc1-bac1869d716d" (UID: "4b251231-3104-491a-9dc1-bac1869d716d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.416999 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4b251231-3104-491a-9dc1-bac1869d716d" (UID: "4b251231-3104-491a-9dc1-bac1869d716d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.492742 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.492776 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.492788 4708 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b251231-3104-491a-9dc1-bac1869d716d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.492802 4708 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4b251231-3104-491a-9dc1-bac1869d716d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.492812 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hsdrn\" (UniqueName: \"kubernetes.io/projected/4b251231-3104-491a-9dc1-bac1869d716d-kube-api-access-hsdrn\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.915647 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" event={"ID":"4b251231-3104-491a-9dc1-bac1869d716d","Type":"ContainerDied","Data":"58d01c124f215a3d213dd3bfc7ba3f76b496047a1ba7883312c730abd7b5b024"} Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.915951 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58d01c124f215a3d213dd3bfc7ba3f76b496047a1ba7883312c730abd7b5b024" Nov 25 06:09:10 crc kubenswrapper[4708]: I1125 06:09:10.916088 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kh7hm" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.012810 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz"] Nov 25 06:09:11 crc kubenswrapper[4708]: E1125 06:09:11.013659 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b251231-3104-491a-9dc1-bac1869d716d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.013678 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b251231-3104-491a-9dc1-bac1869d716d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.013844 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b251231-3104-491a-9dc1-bac1869d716d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.019880 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.021735 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.022246 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.022390 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.022866 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.023244 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.023514 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.024108 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz"] Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117194 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lng7m\" (UniqueName: \"kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117281 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117473 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117621 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117662 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.117985 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219619 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lng7m\" (UniqueName: \"kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219701 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219744 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219771 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219792 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.219870 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.224311 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.224551 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.224691 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.225459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.225612 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.236418 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lng7m\" (UniqueName: \"kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.334589 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.834026 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz"] Nov 25 06:09:11 crc kubenswrapper[4708]: I1125 06:09:11.925678 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" event={"ID":"c88370ee-fc37-4d14-851f-87c2be9e9bf1","Type":"ContainerStarted","Data":"79e3cba8e2545ac5f23a5806887ddc8170eb0e98fabc51ea1c882364956d3f6d"} Nov 25 06:09:12 crc kubenswrapper[4708]: I1125 06:09:12.936188 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" event={"ID":"c88370ee-fc37-4d14-851f-87c2be9e9bf1","Type":"ContainerStarted","Data":"993bb72e492c039b9562eab1a4b88ec9c6e6b41fff8cb7e0375345af62cbc915"} Nov 25 06:09:12 crc kubenswrapper[4708]: I1125 06:09:12.953797 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" podStartSLOduration=2.348045766 podStartE2EDuration="2.953778306s" podCreationTimestamp="2025-11-25 06:09:10 +0000 UTC" firstStartedPulling="2025-11-25 06:09:11.840748539 +0000 UTC m=+1693.249581925" lastFinishedPulling="2025-11-25 06:09:12.44648108 +0000 UTC m=+1693.855314465" observedRunningTime="2025-11-25 06:09:12.953318899 +0000 UTC m=+1694.362152285" watchObservedRunningTime="2025-11-25 06:09:12.953778306 +0000 UTC m=+1694.362611692" Nov 25 06:09:23 crc kubenswrapper[4708]: I1125 06:09:23.893807 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:09:23 crc kubenswrapper[4708]: E1125 06:09:23.894795 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:09:37 crc kubenswrapper[4708]: I1125 06:09:37.894284 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:09:37 crc kubenswrapper[4708]: E1125 06:09:37.895453 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:09:46 crc kubenswrapper[4708]: I1125 06:09:46.278323 4708 generic.go:334] "Generic (PLEG): container finished" podID="c88370ee-fc37-4d14-851f-87c2be9e9bf1" containerID="993bb72e492c039b9562eab1a4b88ec9c6e6b41fff8cb7e0375345af62cbc915" exitCode=0 Nov 25 06:09:46 crc kubenswrapper[4708]: I1125 06:09:46.278413 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" event={"ID":"c88370ee-fc37-4d14-851f-87c2be9e9bf1","Type":"ContainerDied","Data":"993bb72e492c039b9562eab1a4b88ec9c6e6b41fff8cb7e0375345af62cbc915"} Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.620940 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765171 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lng7m\" (UniqueName: \"kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765495 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765552 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765577 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765626 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.765670 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0\") pod \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\" (UID: \"c88370ee-fc37-4d14-851f-87c2be9e9bf1\") " Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.771824 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.772909 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m" (OuterVolumeSpecName: "kube-api-access-lng7m") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "kube-api-access-lng7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.788022 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory" (OuterVolumeSpecName: "inventory") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.788535 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.789244 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.790208 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c88370ee-fc37-4d14-851f-87c2be9e9bf1" (UID: "c88370ee-fc37-4d14-851f-87c2be9e9bf1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868153 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868191 4708 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868207 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868221 4708 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868234 4708 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/c88370ee-fc37-4d14-851f-87c2be9e9bf1-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:47 crc kubenswrapper[4708]: I1125 06:09:47.868248 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lng7m\" (UniqueName: \"kubernetes.io/projected/c88370ee-fc37-4d14-851f-87c2be9e9bf1-kube-api-access-lng7m\") on node \"crc\" DevicePath \"\"" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.297074 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" event={"ID":"c88370ee-fc37-4d14-851f-87c2be9e9bf1","Type":"ContainerDied","Data":"79e3cba8e2545ac5f23a5806887ddc8170eb0e98fabc51ea1c882364956d3f6d"} Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.297132 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79e3cba8e2545ac5f23a5806887ddc8170eb0e98fabc51ea1c882364956d3f6d" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.297169 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.367946 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv"] Nov 25 06:09:48 crc kubenswrapper[4708]: E1125 06:09:48.368312 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c88370ee-fc37-4d14-851f-87c2be9e9bf1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.368332 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c88370ee-fc37-4d14-851f-87c2be9e9bf1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.368542 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c88370ee-fc37-4d14-851f-87c2be9e9bf1" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.369156 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.372252 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.372310 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.372456 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.372554 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.375275 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv"] Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.375413 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.479283 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.479508 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.479875 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gd79\" (UniqueName: \"kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.480038 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.480100 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.582700 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.582769 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.582985 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.583234 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.583388 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gd79\" (UniqueName: \"kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.587933 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.588565 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.589048 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.589480 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.599638 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gd79\" (UniqueName: \"kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-bslnv\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:48 crc kubenswrapper[4708]: I1125 06:09:48.688446 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:09:49 crc kubenswrapper[4708]: I1125 06:09:49.139187 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv"] Nov 25 06:09:49 crc kubenswrapper[4708]: W1125 06:09:49.144509 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb24673ce_c915_4c77_aa4a_83ad370e794e.slice/crio-d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424 WatchSource:0}: Error finding container d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424: Status 404 returned error can't find the container with id d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424 Nov 25 06:09:49 crc kubenswrapper[4708]: I1125 06:09:49.313944 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" event={"ID":"b24673ce-c915-4c77-aa4a-83ad370e794e","Type":"ContainerStarted","Data":"d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424"} Nov 25 06:09:50 crc kubenswrapper[4708]: I1125 06:09:50.325802 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" event={"ID":"b24673ce-c915-4c77-aa4a-83ad370e794e","Type":"ContainerStarted","Data":"e076a5b041d7e5b2ac935e2bcf32d1052113bb0971a9b36f367b54cc6d5b0278"} Nov 25 06:09:50 crc kubenswrapper[4708]: I1125 06:09:50.346837 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" podStartSLOduration=1.807324299 podStartE2EDuration="2.346816892s" podCreationTimestamp="2025-11-25 06:09:48 +0000 UTC" firstStartedPulling="2025-11-25 06:09:49.146960774 +0000 UTC m=+1730.555794160" lastFinishedPulling="2025-11-25 06:09:49.686453367 +0000 UTC m=+1731.095286753" observedRunningTime="2025-11-25 06:09:50.340268228 +0000 UTC m=+1731.749101614" watchObservedRunningTime="2025-11-25 06:09:50.346816892 +0000 UTC m=+1731.755650269" Nov 25 06:09:52 crc kubenswrapper[4708]: I1125 06:09:52.895065 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:09:52 crc kubenswrapper[4708]: E1125 06:09:52.895587 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:10:04 crc kubenswrapper[4708]: I1125 06:10:04.892953 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:10:04 crc kubenswrapper[4708]: E1125 06:10:04.896566 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:10:17 crc kubenswrapper[4708]: I1125 06:10:17.893857 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:10:17 crc kubenswrapper[4708]: E1125 06:10:17.894806 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:10:28 crc kubenswrapper[4708]: I1125 06:10:28.899251 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:10:28 crc kubenswrapper[4708]: E1125 06:10:28.900408 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.273919 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.275801 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.285813 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.454176 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.454269 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.454432 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd2r5\" (UniqueName: \"kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.555346 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.555436 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.555481 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd2r5\" (UniqueName: \"kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.555875 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.555979 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.579774 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd2r5\" (UniqueName: \"kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5\") pod \"certified-operators-44xgx\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:30 crc kubenswrapper[4708]: I1125 06:10:30.606770 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:31 crc kubenswrapper[4708]: I1125 06:10:31.118342 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:31 crc kubenswrapper[4708]: I1125 06:10:31.696091 4708 generic.go:334] "Generic (PLEG): container finished" podID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerID="5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d" exitCode=0 Nov 25 06:10:31 crc kubenswrapper[4708]: I1125 06:10:31.696210 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerDied","Data":"5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d"} Nov 25 06:10:31 crc kubenswrapper[4708]: I1125 06:10:31.696391 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerStarted","Data":"e67b87ff5f53026fe3d5f09b3aaeb47432e1dd72aa48533c15f02ec052a11d62"} Nov 25 06:10:31 crc kubenswrapper[4708]: I1125 06:10:31.698625 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.707219 4708 generic.go:334] "Generic (PLEG): container finished" podID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerID="49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b" exitCode=0 Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.707290 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerDied","Data":"49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b"} Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.872902 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.875104 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.888113 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.911592 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.911686 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv4fz\" (UniqueName: \"kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:32 crc kubenswrapper[4708]: I1125 06:10:32.911846 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.013915 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.013991 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv4fz\" (UniqueName: \"kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.014035 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.014444 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.014500 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.032281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv4fz\" (UniqueName: \"kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz\") pod \"redhat-marketplace-h6295\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.194432 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.640956 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:33 crc kubenswrapper[4708]: W1125 06:10:33.644886 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb61a93f4_f670_42c6_8ff2_98214bd0fc15.slice/crio-8c460db7603af2efdaaf38613cfce19e40cf0517b1d0ed3996d88dab7ddb8dff WatchSource:0}: Error finding container 8c460db7603af2efdaaf38613cfce19e40cf0517b1d0ed3996d88dab7ddb8dff: Status 404 returned error can't find the container with id 8c460db7603af2efdaaf38613cfce19e40cf0517b1d0ed3996d88dab7ddb8dff Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.719868 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerStarted","Data":"8c460db7603af2efdaaf38613cfce19e40cf0517b1d0ed3996d88dab7ddb8dff"} Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.723331 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerStarted","Data":"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03"} Nov 25 06:10:33 crc kubenswrapper[4708]: I1125 06:10:33.748443 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-44xgx" podStartSLOduration=2.275101166 podStartE2EDuration="3.748413979s" podCreationTimestamp="2025-11-25 06:10:30 +0000 UTC" firstStartedPulling="2025-11-25 06:10:31.698354432 +0000 UTC m=+1773.107187819" lastFinishedPulling="2025-11-25 06:10:33.171667246 +0000 UTC m=+1774.580500632" observedRunningTime="2025-11-25 06:10:33.739776428 +0000 UTC m=+1775.148609804" watchObservedRunningTime="2025-11-25 06:10:33.748413979 +0000 UTC m=+1775.157247365" Nov 25 06:10:34 crc kubenswrapper[4708]: I1125 06:10:34.735676 4708 generic.go:334] "Generic (PLEG): container finished" podID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerID="37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f" exitCode=0 Nov 25 06:10:34 crc kubenswrapper[4708]: I1125 06:10:34.735790 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerDied","Data":"37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f"} Nov 25 06:10:35 crc kubenswrapper[4708]: I1125 06:10:35.749002 4708 generic.go:334] "Generic (PLEG): container finished" podID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerID="aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488" exitCode=0 Nov 25 06:10:35 crc kubenswrapper[4708]: I1125 06:10:35.749132 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerDied","Data":"aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488"} Nov 25 06:10:36 crc kubenswrapper[4708]: I1125 06:10:36.762647 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerStarted","Data":"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d"} Nov 25 06:10:36 crc kubenswrapper[4708]: I1125 06:10:36.787922 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h6295" podStartSLOduration=3.292490267 podStartE2EDuration="4.78789419s" podCreationTimestamp="2025-11-25 06:10:32 +0000 UTC" firstStartedPulling="2025-11-25 06:10:34.737678208 +0000 UTC m=+1776.146511594" lastFinishedPulling="2025-11-25 06:10:36.233082131 +0000 UTC m=+1777.641915517" observedRunningTime="2025-11-25 06:10:36.779651102 +0000 UTC m=+1778.188484488" watchObservedRunningTime="2025-11-25 06:10:36.78789419 +0000 UTC m=+1778.196727576" Nov 25 06:10:40 crc kubenswrapper[4708]: I1125 06:10:40.607424 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:40 crc kubenswrapper[4708]: I1125 06:10:40.608339 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:40 crc kubenswrapper[4708]: I1125 06:10:40.645411 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:40 crc kubenswrapper[4708]: I1125 06:10:40.843462 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:40 crc kubenswrapper[4708]: I1125 06:10:40.891281 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:42 crc kubenswrapper[4708]: I1125 06:10:42.822681 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-44xgx" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="registry-server" containerID="cri-o://c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03" gracePeriod=2 Nov 25 06:10:42 crc kubenswrapper[4708]: I1125 06:10:42.894441 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:10:42 crc kubenswrapper[4708]: E1125 06:10:42.894904 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.195492 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.195758 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.207372 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.238165 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.326281 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities\") pod \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.326391 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content\") pod \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.326446 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd2r5\" (UniqueName: \"kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5\") pod \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\" (UID: \"3593ae5c-8fd5-4d34-a495-96b3b41eb10a\") " Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.327263 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities" (OuterVolumeSpecName: "utilities") pod "3593ae5c-8fd5-4d34-a495-96b3b41eb10a" (UID: "3593ae5c-8fd5-4d34-a495-96b3b41eb10a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.333863 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5" (OuterVolumeSpecName: "kube-api-access-vd2r5") pod "3593ae5c-8fd5-4d34-a495-96b3b41eb10a" (UID: "3593ae5c-8fd5-4d34-a495-96b3b41eb10a"). InnerVolumeSpecName "kube-api-access-vd2r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.364607 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3593ae5c-8fd5-4d34-a495-96b3b41eb10a" (UID: "3593ae5c-8fd5-4d34-a495-96b3b41eb10a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.428065 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.428096 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd2r5\" (UniqueName: \"kubernetes.io/projected/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-kube-api-access-vd2r5\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.428110 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3593ae5c-8fd5-4d34-a495-96b3b41eb10a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.835910 4708 generic.go:334] "Generic (PLEG): container finished" podID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerID="c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03" exitCode=0 Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.835981 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerDied","Data":"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03"} Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.836059 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-44xgx" event={"ID":"3593ae5c-8fd5-4d34-a495-96b3b41eb10a","Type":"ContainerDied","Data":"e67b87ff5f53026fe3d5f09b3aaeb47432e1dd72aa48533c15f02ec052a11d62"} Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.836089 4708 scope.go:117] "RemoveContainer" containerID="c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.836005 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-44xgx" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.865277 4708 scope.go:117] "RemoveContainer" containerID="49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.868596 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.874157 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-44xgx"] Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.876339 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.885624 4708 scope.go:117] "RemoveContainer" containerID="5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.915377 4708 scope.go:117] "RemoveContainer" containerID="c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03" Nov 25 06:10:43 crc kubenswrapper[4708]: E1125 06:10:43.916126 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03\": container with ID starting with c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03 not found: ID does not exist" containerID="c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.916177 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03"} err="failed to get container status \"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03\": rpc error: code = NotFound desc = could not find container \"c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03\": container with ID starting with c28985b06154f3ba0b50ff704e100c8bd7ed35b8317e26ba35d095a05fd50d03 not found: ID does not exist" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.916198 4708 scope.go:117] "RemoveContainer" containerID="49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b" Nov 25 06:10:43 crc kubenswrapper[4708]: E1125 06:10:43.917121 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b\": container with ID starting with 49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b not found: ID does not exist" containerID="49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.917178 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b"} err="failed to get container status \"49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b\": rpc error: code = NotFound desc = could not find container \"49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b\": container with ID starting with 49526af2c94fad226d8120f50315c07efe8a28bfec21f18b2aa59cc47089f29b not found: ID does not exist" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.917218 4708 scope.go:117] "RemoveContainer" containerID="5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d" Nov 25 06:10:43 crc kubenswrapper[4708]: E1125 06:10:43.917608 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d\": container with ID starting with 5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d not found: ID does not exist" containerID="5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d" Nov 25 06:10:43 crc kubenswrapper[4708]: I1125 06:10:43.917637 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d"} err="failed to get container status \"5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d\": rpc error: code = NotFound desc = could not find container \"5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d\": container with ID starting with 5348e7c2b3b37333b3426ee19b7c8f327e5cc58446b2a81ebfaf6663db71581d not found: ID does not exist" Nov 25 06:10:44 crc kubenswrapper[4708]: I1125 06:10:44.682145 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:44 crc kubenswrapper[4708]: I1125 06:10:44.903859 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" path="/var/lib/kubelet/pods/3593ae5c-8fd5-4d34-a495-96b3b41eb10a/volumes" Nov 25 06:10:45 crc kubenswrapper[4708]: I1125 06:10:45.854669 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h6295" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="registry-server" containerID="cri-o://4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d" gracePeriod=2 Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.237170 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.297346 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content\") pod \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.298582 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities\") pod \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.298638 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bv4fz\" (UniqueName: \"kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz\") pod \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\" (UID: \"b61a93f4-f670-42c6-8ff2-98214bd0fc15\") " Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.299188 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities" (OuterVolumeSpecName: "utilities") pod "b61a93f4-f670-42c6-8ff2-98214bd0fc15" (UID: "b61a93f4-f670-42c6-8ff2-98214bd0fc15"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.299433 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.304307 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz" (OuterVolumeSpecName: "kube-api-access-bv4fz") pod "b61a93f4-f670-42c6-8ff2-98214bd0fc15" (UID: "b61a93f4-f670-42c6-8ff2-98214bd0fc15"). InnerVolumeSpecName "kube-api-access-bv4fz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.311060 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b61a93f4-f670-42c6-8ff2-98214bd0fc15" (UID: "b61a93f4-f670-42c6-8ff2-98214bd0fc15"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.402146 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b61a93f4-f670-42c6-8ff2-98214bd0fc15-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.402554 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bv4fz\" (UniqueName: \"kubernetes.io/projected/b61a93f4-f670-42c6-8ff2-98214bd0fc15-kube-api-access-bv4fz\") on node \"crc\" DevicePath \"\"" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.863188 4708 generic.go:334] "Generic (PLEG): container finished" podID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerID="4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d" exitCode=0 Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.863250 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6295" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.863259 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerDied","Data":"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d"} Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.863314 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6295" event={"ID":"b61a93f4-f670-42c6-8ff2-98214bd0fc15","Type":"ContainerDied","Data":"8c460db7603af2efdaaf38613cfce19e40cf0517b1d0ed3996d88dab7ddb8dff"} Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.863335 4708 scope.go:117] "RemoveContainer" containerID="4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.879876 4708 scope.go:117] "RemoveContainer" containerID="aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.892595 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.901543 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6295"] Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.912011 4708 scope.go:117] "RemoveContainer" containerID="37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.930741 4708 scope.go:117] "RemoveContainer" containerID="4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d" Nov 25 06:10:46 crc kubenswrapper[4708]: E1125 06:10:46.931247 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d\": container with ID starting with 4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d not found: ID does not exist" containerID="4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.931290 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d"} err="failed to get container status \"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d\": rpc error: code = NotFound desc = could not find container \"4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d\": container with ID starting with 4dbc64ed84e850c77255e7de7a0e77d22c5192d9404d0a65b5318e858526403d not found: ID does not exist" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.931319 4708 scope.go:117] "RemoveContainer" containerID="aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488" Nov 25 06:10:46 crc kubenswrapper[4708]: E1125 06:10:46.931616 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488\": container with ID starting with aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488 not found: ID does not exist" containerID="aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.931638 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488"} err="failed to get container status \"aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488\": rpc error: code = NotFound desc = could not find container \"aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488\": container with ID starting with aa176bd741adc5dcd5dcdb980579b0c2d4c82dcda21d8b7ea4846f88d4785488 not found: ID does not exist" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.931654 4708 scope.go:117] "RemoveContainer" containerID="37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f" Nov 25 06:10:46 crc kubenswrapper[4708]: E1125 06:10:46.931978 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f\": container with ID starting with 37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f not found: ID does not exist" containerID="37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f" Nov 25 06:10:46 crc kubenswrapper[4708]: I1125 06:10:46.932001 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f"} err="failed to get container status \"37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f\": rpc error: code = NotFound desc = could not find container \"37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f\": container with ID starting with 37802ebfb710744ccceda46dfd11b0c0f320dd5207c7605785590006f9ccaa2f not found: ID does not exist" Nov 25 06:10:48 crc kubenswrapper[4708]: I1125 06:10:48.902995 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" path="/var/lib/kubelet/pods/b61a93f4-f670-42c6-8ff2-98214bd0fc15/volumes" Nov 25 06:10:53 crc kubenswrapper[4708]: I1125 06:10:53.893776 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:10:54 crc kubenswrapper[4708]: I1125 06:10:54.934412 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894"} Nov 25 06:13:14 crc kubenswrapper[4708]: I1125 06:13:14.357303 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:13:14 crc kubenswrapper[4708]: I1125 06:13:14.358066 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:13:22 crc kubenswrapper[4708]: I1125 06:13:22.164570 4708 generic.go:334] "Generic (PLEG): container finished" podID="b24673ce-c915-4c77-aa4a-83ad370e794e" containerID="e076a5b041d7e5b2ac935e2bcf32d1052113bb0971a9b36f367b54cc6d5b0278" exitCode=0 Nov 25 06:13:22 crc kubenswrapper[4708]: I1125 06:13:22.164639 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" event={"ID":"b24673ce-c915-4c77-aa4a-83ad370e794e","Type":"ContainerDied","Data":"e076a5b041d7e5b2ac935e2bcf32d1052113bb0971a9b36f367b54cc6d5b0278"} Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.498037 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.696470 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle\") pod \"b24673ce-c915-4c77-aa4a-83ad370e794e\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.696513 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0\") pod \"b24673ce-c915-4c77-aa4a-83ad370e794e\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.696638 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory\") pod \"b24673ce-c915-4c77-aa4a-83ad370e794e\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.696761 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gd79\" (UniqueName: \"kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79\") pod \"b24673ce-c915-4c77-aa4a-83ad370e794e\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.696781 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key\") pod \"b24673ce-c915-4c77-aa4a-83ad370e794e\" (UID: \"b24673ce-c915-4c77-aa4a-83ad370e794e\") " Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.703378 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b24673ce-c915-4c77-aa4a-83ad370e794e" (UID: "b24673ce-c915-4c77-aa4a-83ad370e794e"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.703556 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79" (OuterVolumeSpecName: "kube-api-access-2gd79") pod "b24673ce-c915-4c77-aa4a-83ad370e794e" (UID: "b24673ce-c915-4c77-aa4a-83ad370e794e"). InnerVolumeSpecName "kube-api-access-2gd79". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.721152 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory" (OuterVolumeSpecName: "inventory") pod "b24673ce-c915-4c77-aa4a-83ad370e794e" (UID: "b24673ce-c915-4c77-aa4a-83ad370e794e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.721201 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b24673ce-c915-4c77-aa4a-83ad370e794e" (UID: "b24673ce-c915-4c77-aa4a-83ad370e794e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.721505 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "b24673ce-c915-4c77-aa4a-83ad370e794e" (UID: "b24673ce-c915-4c77-aa4a-83ad370e794e"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.798648 4708 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.798683 4708 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.798695 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.798704 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gd79\" (UniqueName: \"kubernetes.io/projected/b24673ce-c915-4c77-aa4a-83ad370e794e-kube-api-access-2gd79\") on node \"crc\" DevicePath \"\"" Nov 25 06:13:23 crc kubenswrapper[4708]: I1125 06:13:23.798715 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b24673ce-c915-4c77-aa4a-83ad370e794e-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.181236 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" event={"ID":"b24673ce-c915-4c77-aa4a-83ad370e794e","Type":"ContainerDied","Data":"d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424"} Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.181283 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6f755bc8b4056628b4c5b72e330adafad80f1e67eb5570a67da401a136e9424" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.181338 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-bslnv" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260048 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56"] Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260380 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="extract-content" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260398 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="extract-content" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260412 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260420 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260430 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24673ce-c915-4c77-aa4a-83ad370e794e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260437 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24673ce-c915-4c77-aa4a-83ad370e794e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260454 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="extract-utilities" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260461 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="extract-utilities" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260473 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260479 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260486 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="extract-utilities" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260493 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="extract-utilities" Nov 25 06:13:24 crc kubenswrapper[4708]: E1125 06:13:24.260503 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="extract-content" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260509 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="extract-content" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260670 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24673ce-c915-4c77-aa4a-83ad370e794e" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260693 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3593ae5c-8fd5-4d34-a495-96b3b41eb10a" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.260701 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b61a93f4-f670-42c6-8ff2-98214bd0fc15" containerName="registry-server" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.261265 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.264690 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.264777 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.264891 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.265122 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.265199 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.265257 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.265330 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.268788 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56"] Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305631 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305676 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305770 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305806 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305874 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305935 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkf4g\" (UniqueName: \"kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.305977 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.306016 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.306136 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408017 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408104 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408130 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408165 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkf4g\" (UniqueName: \"kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408191 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408231 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408260 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408293 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.408311 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.409780 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.412759 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.412826 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.413404 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.414496 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.414541 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.414784 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.414953 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.421712 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkf4g\" (UniqueName: \"kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g\") pod \"nova-edpm-deployment-openstack-edpm-ipam-wxw56\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:24 crc kubenswrapper[4708]: I1125 06:13:24.582127 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:13:25 crc kubenswrapper[4708]: I1125 06:13:25.028804 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56"] Nov 25 06:13:25 crc kubenswrapper[4708]: I1125 06:13:25.191081 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" event={"ID":"7497674d-bdb1-427d-b4ae-cda53421f90d","Type":"ContainerStarted","Data":"819a85a5f1fc164668daad6ceebac32d068a36c180f46b1965f8c8eaf4098e37"} Nov 25 06:13:26 crc kubenswrapper[4708]: I1125 06:13:26.200338 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" event={"ID":"7497674d-bdb1-427d-b4ae-cda53421f90d","Type":"ContainerStarted","Data":"c6a187c227cc9cc4e1dae0e9bbf6c02e7da3fad4aad4c6d1255845eededa452d"} Nov 25 06:13:26 crc kubenswrapper[4708]: I1125 06:13:26.217677 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" podStartSLOduration=1.624733454 podStartE2EDuration="2.217658849s" podCreationTimestamp="2025-11-25 06:13:24 +0000 UTC" firstStartedPulling="2025-11-25 06:13:25.035338345 +0000 UTC m=+1946.444171731" lastFinishedPulling="2025-11-25 06:13:25.62826374 +0000 UTC m=+1947.037097126" observedRunningTime="2025-11-25 06:13:26.213116516 +0000 UTC m=+1947.621949902" watchObservedRunningTime="2025-11-25 06:13:26.217658849 +0000 UTC m=+1947.626492234" Nov 25 06:13:44 crc kubenswrapper[4708]: I1125 06:13:44.357064 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:13:44 crc kubenswrapper[4708]: I1125 06:13:44.357749 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.357417 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.358150 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.358204 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.358794 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.358849 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894" gracePeriod=600 Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.566732 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894" exitCode=0 Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.566836 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894"} Nov 25 06:14:14 crc kubenswrapper[4708]: I1125 06:14:14.566987 4708 scope.go:117] "RemoveContainer" containerID="5a0a41c22ae727857f7e53479a3b9ae674ca86a4d7e52b28ae32d0df6327cfcf" Nov 25 06:14:15 crc kubenswrapper[4708]: I1125 06:14:15.577241 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197"} Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.141346 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc"] Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.143820 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.145741 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.145936 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.153199 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc"] Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.291672 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.291872 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7g6m\" (UniqueName: \"kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.291954 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.394605 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7g6m\" (UniqueName: \"kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.394696 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.394756 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.395578 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.400651 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.409928 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7g6m\" (UniqueName: \"kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m\") pod \"collect-profiles-29400855-qqmdc\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.468050 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.877239 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc"] Nov 25 06:15:00 crc kubenswrapper[4708]: I1125 06:15:00.920235 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" event={"ID":"c24a9842-5cc0-4e18-8c58-279c982d1a34","Type":"ContainerStarted","Data":"dba3c4335137e4d5de6c022a33e37871a6e869a9998b5a60e717fd49870ecf44"} Nov 25 06:15:01 crc kubenswrapper[4708]: I1125 06:15:01.929933 4708 generic.go:334] "Generic (PLEG): container finished" podID="c24a9842-5cc0-4e18-8c58-279c982d1a34" containerID="211a6060a09526238cc34873ba6d1caaf5e41898cbb3934f1d1521eedaa92de0" exitCode=0 Nov 25 06:15:01 crc kubenswrapper[4708]: I1125 06:15:01.929990 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" event={"ID":"c24a9842-5cc0-4e18-8c58-279c982d1a34","Type":"ContainerDied","Data":"211a6060a09526238cc34873ba6d1caaf5e41898cbb3934f1d1521eedaa92de0"} Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.206901 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.354516 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7g6m\" (UniqueName: \"kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m\") pod \"c24a9842-5cc0-4e18-8c58-279c982d1a34\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.354772 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume\") pod \"c24a9842-5cc0-4e18-8c58-279c982d1a34\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.354816 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume\") pod \"c24a9842-5cc0-4e18-8c58-279c982d1a34\" (UID: \"c24a9842-5cc0-4e18-8c58-279c982d1a34\") " Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.355371 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume" (OuterVolumeSpecName: "config-volume") pod "c24a9842-5cc0-4e18-8c58-279c982d1a34" (UID: "c24a9842-5cc0-4e18-8c58-279c982d1a34"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.359451 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c24a9842-5cc0-4e18-8c58-279c982d1a34" (UID: "c24a9842-5cc0-4e18-8c58-279c982d1a34"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.359745 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m" (OuterVolumeSpecName: "kube-api-access-r7g6m") pod "c24a9842-5cc0-4e18-8c58-279c982d1a34" (UID: "c24a9842-5cc0-4e18-8c58-279c982d1a34"). InnerVolumeSpecName "kube-api-access-r7g6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.457434 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c24a9842-5cc0-4e18-8c58-279c982d1a34-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.457461 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7g6m\" (UniqueName: \"kubernetes.io/projected/c24a9842-5cc0-4e18-8c58-279c982d1a34-kube-api-access-r7g6m\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.457473 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c24a9842-5cc0-4e18-8c58-279c982d1a34-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.947230 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" event={"ID":"c24a9842-5cc0-4e18-8c58-279c982d1a34","Type":"ContainerDied","Data":"dba3c4335137e4d5de6c022a33e37871a6e869a9998b5a60e717fd49870ecf44"} Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.947493 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dba3c4335137e4d5de6c022a33e37871a6e869a9998b5a60e717fd49870ecf44" Nov 25 06:15:03 crc kubenswrapper[4708]: I1125 06:15:03.947272 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400855-qqmdc" Nov 25 06:15:04 crc kubenswrapper[4708]: I1125 06:15:04.280400 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r"] Nov 25 06:15:04 crc kubenswrapper[4708]: I1125 06:15:04.288553 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400810-qcn4r"] Nov 25 06:15:04 crc kubenswrapper[4708]: I1125 06:15:04.902627 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c35d3991-35e6-4cd7-a3c9-11513018d1fd" path="/var/lib/kubelet/pods/c35d3991-35e6-4cd7-a3c9-11513018d1fd/volumes" Nov 25 06:15:16 crc kubenswrapper[4708]: I1125 06:15:16.054399 4708 generic.go:334] "Generic (PLEG): container finished" podID="7497674d-bdb1-427d-b4ae-cda53421f90d" containerID="c6a187c227cc9cc4e1dae0e9bbf6c02e7da3fad4aad4c6d1255845eededa452d" exitCode=0 Nov 25 06:15:16 crc kubenswrapper[4708]: I1125 06:15:16.055009 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" event={"ID":"7497674d-bdb1-427d-b4ae-cda53421f90d","Type":"ContainerDied","Data":"c6a187c227cc9cc4e1dae0e9bbf6c02e7da3fad4aad4c6d1255845eededa452d"} Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.407588 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534502 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534572 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534592 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534648 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkf4g\" (UniqueName: \"kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534679 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534725 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534750 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534786 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.534803 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0\") pod \"7497674d-bdb1-427d-b4ae-cda53421f90d\" (UID: \"7497674d-bdb1-427d-b4ae-cda53421f90d\") " Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.541036 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g" (OuterVolumeSpecName: "kube-api-access-hkf4g") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "kube-api-access-hkf4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.542306 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.561676 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.564072 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.564895 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.566880 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.567211 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.568024 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.569990 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory" (OuterVolumeSpecName: "inventory") pod "7497674d-bdb1-427d-b4ae-cda53421f90d" (UID: "7497674d-bdb1-427d-b4ae-cda53421f90d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637680 4708 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637720 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637731 4708 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637745 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkf4g\" (UniqueName: \"kubernetes.io/projected/7497674d-bdb1-427d-b4ae-cda53421f90d-kube-api-access-hkf4g\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637757 4708 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637769 4708 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637780 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637790 4708 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:17 crc kubenswrapper[4708]: I1125 06:15:17.637802 4708 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/7497674d-bdb1-427d-b4ae-cda53421f90d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.077692 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" event={"ID":"7497674d-bdb1-427d-b4ae-cda53421f90d","Type":"ContainerDied","Data":"819a85a5f1fc164668daad6ceebac32d068a36c180f46b1965f8c8eaf4098e37"} Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.077737 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="819a85a5f1fc164668daad6ceebac32d068a36c180f46b1965f8c8eaf4098e37" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.077763 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-wxw56" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.149256 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q"] Nov 25 06:15:18 crc kubenswrapper[4708]: E1125 06:15:18.149723 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c24a9842-5cc0-4e18-8c58-279c982d1a34" containerName="collect-profiles" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.149742 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c24a9842-5cc0-4e18-8c58-279c982d1a34" containerName="collect-profiles" Nov 25 06:15:18 crc kubenswrapper[4708]: E1125 06:15:18.149785 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7497674d-bdb1-427d-b4ae-cda53421f90d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.149792 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="7497674d-bdb1-427d-b4ae-cda53421f90d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.150003 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c24a9842-5cc0-4e18-8c58-279c982d1a34" containerName="collect-profiles" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.150020 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="7497674d-bdb1-427d-b4ae-cda53421f90d" containerName="nova-edpm-deployment-openstack-edpm-ipam" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.150698 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.152813 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.153340 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.153474 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.153647 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.153779 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jtq52" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.165913 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q"] Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.250101 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhgxh\" (UniqueName: \"kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.250485 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.250748 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.250986 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.251245 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.251503 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.251633 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353466 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353542 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353583 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353638 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353658 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353706 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhgxh\" (UniqueName: \"kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.353731 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.357979 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.358281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.358930 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.359281 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.359369 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.359368 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.369249 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhgxh\" (UniqueName: \"kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.464828 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:15:18 crc kubenswrapper[4708]: I1125 06:15:18.946141 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q"] Nov 25 06:15:19 crc kubenswrapper[4708]: I1125 06:15:19.089754 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" event={"ID":"180c2306-33c0-4aed-93f8-3c99603f678b","Type":"ContainerStarted","Data":"50fe97757e783ba2578cbe20e475aafda378d1a811883baf5d813557145dbed6"} Nov 25 06:15:20 crc kubenswrapper[4708]: I1125 06:15:20.099073 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" event={"ID":"180c2306-33c0-4aed-93f8-3c99603f678b","Type":"ContainerStarted","Data":"18b67e93adcfb01bac554a56e55c8fde8409bc10c6eb7e0d05408d9a43b007c4"} Nov 25 06:15:20 crc kubenswrapper[4708]: I1125 06:15:20.116402 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" podStartSLOduration=1.5235360039999999 podStartE2EDuration="2.116384193s" podCreationTimestamp="2025-11-25 06:15:18 +0000 UTC" firstStartedPulling="2025-11-25 06:15:18.953589487 +0000 UTC m=+2060.362422873" lastFinishedPulling="2025-11-25 06:15:19.546437676 +0000 UTC m=+2060.955271062" observedRunningTime="2025-11-25 06:15:20.115325356 +0000 UTC m=+2061.524158743" watchObservedRunningTime="2025-11-25 06:15:20.116384193 +0000 UTC m=+2061.525217579" Nov 25 06:16:01 crc kubenswrapper[4708]: I1125 06:16:01.853094 4708 scope.go:117] "RemoveContainer" containerID="35d229d1a04c1aa30ecf7da1cf037f3f78d5e52658ddc23b31ba457a7eeadcad" Nov 25 06:16:14 crc kubenswrapper[4708]: I1125 06:16:14.356967 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:16:14 crc kubenswrapper[4708]: I1125 06:16:14.357506 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:16:44 crc kubenswrapper[4708]: I1125 06:16:44.356879 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:16:44 crc kubenswrapper[4708]: I1125 06:16:44.357216 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:16:53 crc kubenswrapper[4708]: I1125 06:16:53.789573 4708 generic.go:334] "Generic (PLEG): container finished" podID="180c2306-33c0-4aed-93f8-3c99603f678b" containerID="18b67e93adcfb01bac554a56e55c8fde8409bc10c6eb7e0d05408d9a43b007c4" exitCode=0 Nov 25 06:16:53 crc kubenswrapper[4708]: I1125 06:16:53.789623 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" event={"ID":"180c2306-33c0-4aed-93f8-3c99603f678b","Type":"ContainerDied","Data":"18b67e93adcfb01bac554a56e55c8fde8409bc10c6eb7e0d05408d9a43b007c4"} Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.144993 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.323846 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324358 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhgxh\" (UniqueName: \"kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324483 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324587 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324690 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324761 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.324823 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0\") pod \"180c2306-33c0-4aed-93f8-3c99603f678b\" (UID: \"180c2306-33c0-4aed-93f8-3c99603f678b\") " Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.335650 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh" (OuterVolumeSpecName: "kube-api-access-fhgxh") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "kube-api-access-fhgxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.335676 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.351230 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.352548 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.353611 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory" (OuterVolumeSpecName: "inventory") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.354292 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.357809 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "180c2306-33c0-4aed-93f8-3c99603f678b" (UID: "180c2306-33c0-4aed-93f8-3c99603f678b"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427855 4708 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427889 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427901 4708 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427912 4708 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427921 4708 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427932 4708 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/180c2306-33c0-4aed-93f8-3c99603f678b-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.427943 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhgxh\" (UniqueName: \"kubernetes.io/projected/180c2306-33c0-4aed-93f8-3c99603f678b-kube-api-access-fhgxh\") on node \"crc\" DevicePath \"\"" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.809355 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" event={"ID":"180c2306-33c0-4aed-93f8-3c99603f678b","Type":"ContainerDied","Data":"50fe97757e783ba2578cbe20e475aafda378d1a811883baf5d813557145dbed6"} Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.809416 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50fe97757e783ba2578cbe20e475aafda378d1a811883baf5d813557145dbed6" Nov 25 06:16:55 crc kubenswrapper[4708]: I1125 06:16:55.809593 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.473407 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:03 crc kubenswrapper[4708]: E1125 06:17:03.474348 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="180c2306-33c0-4aed-93f8-3c99603f678b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.474364 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="180c2306-33c0-4aed-93f8-3c99603f678b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.474598 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="180c2306-33c0-4aed-93f8-3c99603f678b" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.475983 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.489180 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.573685 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.574081 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.574301 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fnb7\" (UniqueName: \"kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.676016 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fnb7\" (UniqueName: \"kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.676109 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.676189 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.676700 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.676757 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.694895 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fnb7\" (UniqueName: \"kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7\") pod \"community-operators-h8qb5\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:03 crc kubenswrapper[4708]: I1125 06:17:03.792721 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:04 crc kubenswrapper[4708]: I1125 06:17:04.309295 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:04 crc kubenswrapper[4708]: E1125 06:17:04.704783 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a505f2f_4ef0_42e9_bf0a_3abda684b108.slice/crio-87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:17:04 crc kubenswrapper[4708]: I1125 06:17:04.887319 4708 generic.go:334] "Generic (PLEG): container finished" podID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerID="87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8" exitCode=0 Nov 25 06:17:04 crc kubenswrapper[4708]: I1125 06:17:04.887366 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerDied","Data":"87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8"} Nov 25 06:17:04 crc kubenswrapper[4708]: I1125 06:17:04.887415 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerStarted","Data":"c624ef26da73a7a13b80c5a225bb0b4b00425275911e60b986d661b09f5fdb4a"} Nov 25 06:17:04 crc kubenswrapper[4708]: I1125 06:17:04.889409 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:17:05 crc kubenswrapper[4708]: I1125 06:17:05.897797 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerStarted","Data":"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae"} Nov 25 06:17:06 crc kubenswrapper[4708]: I1125 06:17:06.920134 4708 generic.go:334] "Generic (PLEG): container finished" podID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerID="a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae" exitCode=0 Nov 25 06:17:06 crc kubenswrapper[4708]: I1125 06:17:06.935053 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerDied","Data":"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae"} Nov 25 06:17:07 crc kubenswrapper[4708]: I1125 06:17:07.950734 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerStarted","Data":"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8"} Nov 25 06:17:07 crc kubenswrapper[4708]: I1125 06:17:07.980654 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h8qb5" podStartSLOduration=2.462316062 podStartE2EDuration="4.980635382s" podCreationTimestamp="2025-11-25 06:17:03 +0000 UTC" firstStartedPulling="2025-11-25 06:17:04.889131444 +0000 UTC m=+2166.297964830" lastFinishedPulling="2025-11-25 06:17:07.407450765 +0000 UTC m=+2168.816284150" observedRunningTime="2025-11-25 06:17:07.976496439 +0000 UTC m=+2169.385329826" watchObservedRunningTime="2025-11-25 06:17:07.980635382 +0000 UTC m=+2169.389468767" Nov 25 06:17:13 crc kubenswrapper[4708]: I1125 06:17:13.793359 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:13 crc kubenswrapper[4708]: I1125 06:17:13.794274 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:13 crc kubenswrapper[4708]: I1125 06:17:13.828719 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.035084 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.071377 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.357330 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.357377 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.357415 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.358052 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:17:14 crc kubenswrapper[4708]: I1125 06:17:14.358104 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" gracePeriod=600 Nov 25 06:17:14 crc kubenswrapper[4708]: E1125 06:17:14.474488 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:17:15 crc kubenswrapper[4708]: I1125 06:17:15.009239 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" exitCode=0 Nov 25 06:17:15 crc kubenswrapper[4708]: I1125 06:17:15.009313 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197"} Nov 25 06:17:15 crc kubenswrapper[4708]: I1125 06:17:15.009686 4708 scope.go:117] "RemoveContainer" containerID="9b3738cc802a69ef566340a11d001a816b7af297272381f57416c02e39467894" Nov 25 06:17:15 crc kubenswrapper[4708]: I1125 06:17:15.011478 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:17:15 crc kubenswrapper[4708]: E1125 06:17:15.012113 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.021030 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h8qb5" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="registry-server" containerID="cri-o://498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8" gracePeriod=2 Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.417193 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.458890 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities\") pod \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.459375 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content\") pod \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.459499 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fnb7\" (UniqueName: \"kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7\") pod \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\" (UID: \"6a505f2f-4ef0-42e9-bf0a-3abda684b108\") " Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.459958 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities" (OuterVolumeSpecName: "utilities") pod "6a505f2f-4ef0-42e9-bf0a-3abda684b108" (UID: "6a505f2f-4ef0-42e9-bf0a-3abda684b108"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.460172 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.465205 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7" (OuterVolumeSpecName: "kube-api-access-4fnb7") pod "6a505f2f-4ef0-42e9-bf0a-3abda684b108" (UID: "6a505f2f-4ef0-42e9-bf0a-3abda684b108"). InnerVolumeSpecName "kube-api-access-4fnb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.501550 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6a505f2f-4ef0-42e9-bf0a-3abda684b108" (UID: "6a505f2f-4ef0-42e9-bf0a-3abda684b108"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.562955 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6a505f2f-4ef0-42e9-bf0a-3abda684b108-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:17:16 crc kubenswrapper[4708]: I1125 06:17:16.562987 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fnb7\" (UniqueName: \"kubernetes.io/projected/6a505f2f-4ef0-42e9-bf0a-3abda684b108-kube-api-access-4fnb7\") on node \"crc\" DevicePath \"\"" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.029391 4708 generic.go:334] "Generic (PLEG): container finished" podID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerID="498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8" exitCode=0 Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.029431 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerDied","Data":"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8"} Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.029456 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h8qb5" event={"ID":"6a505f2f-4ef0-42e9-bf0a-3abda684b108","Type":"ContainerDied","Data":"c624ef26da73a7a13b80c5a225bb0b4b00425275911e60b986d661b09f5fdb4a"} Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.029476 4708 scope.go:117] "RemoveContainer" containerID="498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.029590 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h8qb5" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.048313 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.051559 4708 scope.go:117] "RemoveContainer" containerID="a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.057653 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h8qb5"] Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.067016 4708 scope.go:117] "RemoveContainer" containerID="87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.099298 4708 scope.go:117] "RemoveContainer" containerID="498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8" Nov 25 06:17:17 crc kubenswrapper[4708]: E1125 06:17:17.099698 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8\": container with ID starting with 498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8 not found: ID does not exist" containerID="498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.099736 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8"} err="failed to get container status \"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8\": rpc error: code = NotFound desc = could not find container \"498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8\": container with ID starting with 498953210efd8214553c1fb330fe4c4c4da8aea974546702ffb86b518dc2f9a8 not found: ID does not exist" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.099763 4708 scope.go:117] "RemoveContainer" containerID="a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae" Nov 25 06:17:17 crc kubenswrapper[4708]: E1125 06:17:17.100085 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae\": container with ID starting with a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae not found: ID does not exist" containerID="a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.100137 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae"} err="failed to get container status \"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae\": rpc error: code = NotFound desc = could not find container \"a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae\": container with ID starting with a2c2c581f9862f444d7d927a88e6774255ba8c07f50c85aa974151830568edae not found: ID does not exist" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.100153 4708 scope.go:117] "RemoveContainer" containerID="87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8" Nov 25 06:17:17 crc kubenswrapper[4708]: E1125 06:17:17.100504 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8\": container with ID starting with 87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8 not found: ID does not exist" containerID="87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8" Nov 25 06:17:17 crc kubenswrapper[4708]: I1125 06:17:17.100549 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8"} err="failed to get container status \"87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8\": rpc error: code = NotFound desc = could not find container \"87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8\": container with ID starting with 87ccd2c17d7674df32fef6413d810bada0bb6e6aa675e147f212c209f3b69db8 not found: ID does not exist" Nov 25 06:17:18 crc kubenswrapper[4708]: I1125 06:17:18.935454 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" path="/var/lib/kubelet/pods/6a505f2f-4ef0-42e9-bf0a-3abda684b108/volumes" Nov 25 06:17:28 crc kubenswrapper[4708]: I1125 06:17:28.899054 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:17:28 crc kubenswrapper[4708]: E1125 06:17:28.899821 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:17:40 crc kubenswrapper[4708]: I1125 06:17:40.893208 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:17:40 crc kubenswrapper[4708]: E1125 06:17:40.894538 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.695251 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 06:17:42 crc kubenswrapper[4708]: E1125 06:17:42.695956 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="extract-content" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.695971 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="extract-content" Nov 25 06:17:42 crc kubenswrapper[4708]: E1125 06:17:42.696010 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="extract-utilities" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.696017 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="extract-utilities" Nov 25 06:17:42 crc kubenswrapper[4708]: E1125 06:17:42.696043 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="registry-server" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.696049 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="registry-server" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.696219 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a505f2f-4ef0-42e9-bf0a-3abda684b108" containerName="registry-server" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.696869 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.698470 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.698829 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5cn88" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.698993 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.699128 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.709717 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825144 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825215 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825233 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825269 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825294 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825310 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcc59\" (UniqueName: \"kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825370 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825414 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.825431 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.926940 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927019 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927048 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927088 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927121 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927138 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927167 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927189 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927208 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcc59\" (UniqueName: \"kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927280 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.927922 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.928509 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.929231 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.929432 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.934628 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.934920 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.935488 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.941991 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcc59\" (UniqueName: \"kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:42 crc kubenswrapper[4708]: I1125 06:17:42.950641 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " pod="openstack/tempest-tests-tempest" Nov 25 06:17:43 crc kubenswrapper[4708]: I1125 06:17:43.012707 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 06:17:43 crc kubenswrapper[4708]: I1125 06:17:43.407883 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Nov 25 06:17:44 crc kubenswrapper[4708]: I1125 06:17:44.238536 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"695e5c80-15cf-4dc5-b286-cb10e12e9fee","Type":"ContainerStarted","Data":"0e396f85a4bd65f1b4307ef2aba254d2728c79adb2b3b842d3394b51f87e2428"} Nov 25 06:17:51 crc kubenswrapper[4708]: I1125 06:17:51.893736 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:17:51 crc kubenswrapper[4708]: E1125 06:17:51.894419 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:17:58 crc kubenswrapper[4708]: I1125 06:17:58.357105 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"695e5c80-15cf-4dc5-b286-cb10e12e9fee","Type":"ContainerStarted","Data":"9b53254b7e4efb4305fe6599eb57396df1ac06885e702b8adeaef13814ed51de"} Nov 25 06:17:58 crc kubenswrapper[4708]: I1125 06:17:58.375262 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=2.9784393160000002 podStartE2EDuration="17.375248263s" podCreationTimestamp="2025-11-25 06:17:41 +0000 UTC" firstStartedPulling="2025-11-25 06:17:43.408460752 +0000 UTC m=+2204.817294138" lastFinishedPulling="2025-11-25 06:17:57.805269699 +0000 UTC m=+2219.214103085" observedRunningTime="2025-11-25 06:17:58.369039078 +0000 UTC m=+2219.777872464" watchObservedRunningTime="2025-11-25 06:17:58.375248263 +0000 UTC m=+2219.784081649" Nov 25 06:18:04 crc kubenswrapper[4708]: I1125 06:18:04.894052 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:18:04 crc kubenswrapper[4708]: E1125 06:18:04.894898 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:18:12 crc kubenswrapper[4708]: I1125 06:18:12.958120 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:12 crc kubenswrapper[4708]: I1125 06:18:12.961864 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:12 crc kubenswrapper[4708]: I1125 06:18:12.969384 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.004260 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqs4j\" (UniqueName: \"kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.004641 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.004909 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.107260 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.107635 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqs4j\" (UniqueName: \"kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.107671 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.107771 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.108166 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.125912 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqs4j\" (UniqueName: \"kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j\") pod \"redhat-operators-v6f6f\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.297108 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:13 crc kubenswrapper[4708]: W1125 06:18:13.716153 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5ef8dbd_069d_4fe1_b991_a3bbb617f300.slice/crio-3407b2dc5203a47da53948d2a970c5b9fcb3d1309ec42bb76b2894c81bc0fed9 WatchSource:0}: Error finding container 3407b2dc5203a47da53948d2a970c5b9fcb3d1309ec42bb76b2894c81bc0fed9: Status 404 returned error can't find the container with id 3407b2dc5203a47da53948d2a970c5b9fcb3d1309ec42bb76b2894c81bc0fed9 Nov 25 06:18:13 crc kubenswrapper[4708]: I1125 06:18:13.722691 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:14 crc kubenswrapper[4708]: I1125 06:18:14.496731 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerID="fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053" exitCode=0 Nov 25 06:18:14 crc kubenswrapper[4708]: I1125 06:18:14.496772 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerDied","Data":"fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053"} Nov 25 06:18:14 crc kubenswrapper[4708]: I1125 06:18:14.497119 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerStarted","Data":"3407b2dc5203a47da53948d2a970c5b9fcb3d1309ec42bb76b2894c81bc0fed9"} Nov 25 06:18:16 crc kubenswrapper[4708]: E1125 06:18:16.291074 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5ef8dbd_069d_4fe1_b991_a3bbb617f300.slice/crio-237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:18:16 crc kubenswrapper[4708]: I1125 06:18:16.518535 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerStarted","Data":"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a"} Nov 25 06:18:18 crc kubenswrapper[4708]: I1125 06:18:18.537646 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerID="237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a" exitCode=0 Nov 25 06:18:18 crc kubenswrapper[4708]: I1125 06:18:18.537707 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerDied","Data":"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a"} Nov 25 06:18:19 crc kubenswrapper[4708]: I1125 06:18:19.553585 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerStarted","Data":"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418"} Nov 25 06:18:19 crc kubenswrapper[4708]: I1125 06:18:19.574648 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v6f6f" podStartSLOduration=3.016913372 podStartE2EDuration="7.574632875s" podCreationTimestamp="2025-11-25 06:18:12 +0000 UTC" firstStartedPulling="2025-11-25 06:18:14.500004083 +0000 UTC m=+2235.908837470" lastFinishedPulling="2025-11-25 06:18:19.057723586 +0000 UTC m=+2240.466556973" observedRunningTime="2025-11-25 06:18:19.567132496 +0000 UTC m=+2240.975965882" watchObservedRunningTime="2025-11-25 06:18:19.574632875 +0000 UTC m=+2240.983466261" Nov 25 06:18:19 crc kubenswrapper[4708]: I1125 06:18:19.893025 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:18:19 crc kubenswrapper[4708]: E1125 06:18:19.893236 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:18:23 crc kubenswrapper[4708]: I1125 06:18:23.298709 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:23 crc kubenswrapper[4708]: I1125 06:18:23.299447 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:24 crc kubenswrapper[4708]: I1125 06:18:24.344620 4708 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v6f6f" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="registry-server" probeResult="failure" output=< Nov 25 06:18:24 crc kubenswrapper[4708]: timeout: failed to connect service ":50051" within 1s Nov 25 06:18:24 crc kubenswrapper[4708]: > Nov 25 06:18:33 crc kubenswrapper[4708]: I1125 06:18:33.341938 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:33 crc kubenswrapper[4708]: I1125 06:18:33.382947 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:33 crc kubenswrapper[4708]: I1125 06:18:33.570950 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:34 crc kubenswrapper[4708]: I1125 06:18:34.677675 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v6f6f" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="registry-server" containerID="cri-o://7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418" gracePeriod=2 Nov 25 06:18:34 crc kubenswrapper[4708]: I1125 06:18:34.894592 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:18:34 crc kubenswrapper[4708]: E1125 06:18:34.894791 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.099671 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.176787 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqs4j\" (UniqueName: \"kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j\") pod \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.177100 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities\") pod \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.177208 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content\") pod \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\" (UID: \"c5ef8dbd-069d-4fe1-b991-a3bbb617f300\") " Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.177688 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities" (OuterVolumeSpecName: "utilities") pod "c5ef8dbd-069d-4fe1-b991-a3bbb617f300" (UID: "c5ef8dbd-069d-4fe1-b991-a3bbb617f300"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.182328 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j" (OuterVolumeSpecName: "kube-api-access-pqs4j") pod "c5ef8dbd-069d-4fe1-b991-a3bbb617f300" (UID: "c5ef8dbd-069d-4fe1-b991-a3bbb617f300"). InnerVolumeSpecName "kube-api-access-pqs4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.246732 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5ef8dbd-069d-4fe1-b991-a3bbb617f300" (UID: "c5ef8dbd-069d-4fe1-b991-a3bbb617f300"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.279197 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.279233 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.279247 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqs4j\" (UniqueName: \"kubernetes.io/projected/c5ef8dbd-069d-4fe1-b991-a3bbb617f300-kube-api-access-pqs4j\") on node \"crc\" DevicePath \"\"" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.694259 4708 generic.go:334] "Generic (PLEG): container finished" podID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerID="7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418" exitCode=0 Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.694310 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerDied","Data":"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418"} Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.694351 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v6f6f" event={"ID":"c5ef8dbd-069d-4fe1-b991-a3bbb617f300","Type":"ContainerDied","Data":"3407b2dc5203a47da53948d2a970c5b9fcb3d1309ec42bb76b2894c81bc0fed9"} Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.694371 4708 scope.go:117] "RemoveContainer" containerID="7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.694370 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v6f6f" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.719634 4708 scope.go:117] "RemoveContainer" containerID="237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.728731 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.734742 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v6f6f"] Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.739290 4708 scope.go:117] "RemoveContainer" containerID="fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.773968 4708 scope.go:117] "RemoveContainer" containerID="7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418" Nov 25 06:18:35 crc kubenswrapper[4708]: E1125 06:18:35.774323 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418\": container with ID starting with 7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418 not found: ID does not exist" containerID="7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.774355 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418"} err="failed to get container status \"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418\": rpc error: code = NotFound desc = could not find container \"7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418\": container with ID starting with 7a75efe7c3f440a1c65083509fbcf784446d3fc4dee0ae54658b2589479cd418 not found: ID does not exist" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.774376 4708 scope.go:117] "RemoveContainer" containerID="237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a" Nov 25 06:18:35 crc kubenswrapper[4708]: E1125 06:18:35.774699 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a\": container with ID starting with 237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a not found: ID does not exist" containerID="237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.774722 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a"} err="failed to get container status \"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a\": rpc error: code = NotFound desc = could not find container \"237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a\": container with ID starting with 237ec114d0aaada0672df42fad6e82d28250ef46973bd4fabab9ca7bf6f7b10a not found: ID does not exist" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.774736 4708 scope.go:117] "RemoveContainer" containerID="fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053" Nov 25 06:18:35 crc kubenswrapper[4708]: E1125 06:18:35.774952 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053\": container with ID starting with fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053 not found: ID does not exist" containerID="fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053" Nov 25 06:18:35 crc kubenswrapper[4708]: I1125 06:18:35.774984 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053"} err="failed to get container status \"fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053\": rpc error: code = NotFound desc = could not find container \"fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053\": container with ID starting with fc2e3ccf4a4cc7859bdb7a72136a3f4d3aa11e2c31699600313f8be782247053 not found: ID does not exist" Nov 25 06:18:36 crc kubenswrapper[4708]: I1125 06:18:36.901658 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" path="/var/lib/kubelet/pods/c5ef8dbd-069d-4fe1-b991-a3bbb617f300/volumes" Nov 25 06:18:48 crc kubenswrapper[4708]: I1125 06:18:48.899913 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:18:48 crc kubenswrapper[4708]: E1125 06:18:48.900511 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:18:59 crc kubenswrapper[4708]: I1125 06:18:59.893369 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:18:59 crc kubenswrapper[4708]: E1125 06:18:59.894185 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:19:11 crc kubenswrapper[4708]: I1125 06:19:11.892891 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:19:11 crc kubenswrapper[4708]: E1125 06:19:11.893863 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:19:26 crc kubenswrapper[4708]: I1125 06:19:26.893426 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:19:26 crc kubenswrapper[4708]: E1125 06:19:26.894278 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:19:38 crc kubenswrapper[4708]: I1125 06:19:38.898688 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:19:38 crc kubenswrapper[4708]: E1125 06:19:38.900232 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:19:50 crc kubenswrapper[4708]: I1125 06:19:50.893828 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:19:50 crc kubenswrapper[4708]: E1125 06:19:50.895876 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:20:04 crc kubenswrapper[4708]: I1125 06:20:04.893132 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:20:04 crc kubenswrapper[4708]: E1125 06:20:04.894928 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:20:15 crc kubenswrapper[4708]: I1125 06:20:15.893509 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:20:15 crc kubenswrapper[4708]: E1125 06:20:15.894360 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:20:29 crc kubenswrapper[4708]: I1125 06:20:29.894210 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:20:29 crc kubenswrapper[4708]: E1125 06:20:29.895374 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:20:43 crc kubenswrapper[4708]: I1125 06:20:43.893856 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:20:43 crc kubenswrapper[4708]: E1125 06:20:43.894876 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:20:56 crc kubenswrapper[4708]: I1125 06:20:56.893778 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:20:56 crc kubenswrapper[4708]: E1125 06:20:56.894566 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.570507 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:03 crc kubenswrapper[4708]: E1125 06:21:03.571942 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="extract-content" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.571965 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="extract-content" Nov 25 06:21:03 crc kubenswrapper[4708]: E1125 06:21:03.572001 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="extract-utilities" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.572011 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="extract-utilities" Nov 25 06:21:03 crc kubenswrapper[4708]: E1125 06:21:03.572036 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="registry-server" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.572042 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="registry-server" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.572264 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5ef8dbd-069d-4fe1-b991-a3bbb617f300" containerName="registry-server" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.574164 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.590512 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.769936 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.770126 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.770235 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr8sg\" (UniqueName: \"kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.871582 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.871649 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr8sg\" (UniqueName: \"kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.871774 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.872088 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.872141 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.888406 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr8sg\" (UniqueName: \"kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg\") pod \"certified-operators-psxsl\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:03 crc kubenswrapper[4708]: I1125 06:21:03.907181 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:04 crc kubenswrapper[4708]: I1125 06:21:04.390024 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:04 crc kubenswrapper[4708]: W1125 06:21:04.392471 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61729bed_368a_498d_b6f8_fcadf63835fc.slice/crio-d0c3645f2b8c98188f72d5180976da69d8572d22dab00d789f868d3e0b075935 WatchSource:0}: Error finding container d0c3645f2b8c98188f72d5180976da69d8572d22dab00d789f868d3e0b075935: Status 404 returned error can't find the container with id d0c3645f2b8c98188f72d5180976da69d8572d22dab00d789f868d3e0b075935 Nov 25 06:21:04 crc kubenswrapper[4708]: I1125 06:21:04.816415 4708 generic.go:334] "Generic (PLEG): container finished" podID="61729bed-368a-498d-b6f8-fcadf63835fc" containerID="362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb" exitCode=0 Nov 25 06:21:04 crc kubenswrapper[4708]: I1125 06:21:04.816476 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerDied","Data":"362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb"} Nov 25 06:21:04 crc kubenswrapper[4708]: I1125 06:21:04.816546 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerStarted","Data":"d0c3645f2b8c98188f72d5180976da69d8572d22dab00d789f868d3e0b075935"} Nov 25 06:21:05 crc kubenswrapper[4708]: I1125 06:21:05.827374 4708 generic.go:334] "Generic (PLEG): container finished" podID="61729bed-368a-498d-b6f8-fcadf63835fc" containerID="2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239" exitCode=0 Nov 25 06:21:05 crc kubenswrapper[4708]: I1125 06:21:05.827628 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerDied","Data":"2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239"} Nov 25 06:21:06 crc kubenswrapper[4708]: I1125 06:21:06.845345 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerStarted","Data":"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c"} Nov 25 06:21:06 crc kubenswrapper[4708]: I1125 06:21:06.869809 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-psxsl" podStartSLOduration=2.347546325 podStartE2EDuration="3.869789662s" podCreationTimestamp="2025-11-25 06:21:03 +0000 UTC" firstStartedPulling="2025-11-25 06:21:04.818834419 +0000 UTC m=+2406.227667805" lastFinishedPulling="2025-11-25 06:21:06.341077756 +0000 UTC m=+2407.749911142" observedRunningTime="2025-11-25 06:21:06.861949733 +0000 UTC m=+2408.270783120" watchObservedRunningTime="2025-11-25 06:21:06.869789662 +0000 UTC m=+2408.278623048" Nov 25 06:21:11 crc kubenswrapper[4708]: I1125 06:21:11.893495 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:21:11 crc kubenswrapper[4708]: E1125 06:21:11.894325 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:21:13 crc kubenswrapper[4708]: I1125 06:21:13.908082 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:13 crc kubenswrapper[4708]: I1125 06:21:13.908137 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:13 crc kubenswrapper[4708]: I1125 06:21:13.952936 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:14 crc kubenswrapper[4708]: I1125 06:21:14.950053 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:14 crc kubenswrapper[4708]: I1125 06:21:14.989764 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:16 crc kubenswrapper[4708]: I1125 06:21:16.925499 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-psxsl" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="registry-server" containerID="cri-o://433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c" gracePeriod=2 Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.291577 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.438234 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities\") pod \"61729bed-368a-498d-b6f8-fcadf63835fc\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.438476 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content\") pod \"61729bed-368a-498d-b6f8-fcadf63835fc\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.438541 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr8sg\" (UniqueName: \"kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg\") pod \"61729bed-368a-498d-b6f8-fcadf63835fc\" (UID: \"61729bed-368a-498d-b6f8-fcadf63835fc\") " Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.439016 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities" (OuterVolumeSpecName: "utilities") pod "61729bed-368a-498d-b6f8-fcadf63835fc" (UID: "61729bed-368a-498d-b6f8-fcadf63835fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.443377 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg" (OuterVolumeSpecName: "kube-api-access-kr8sg") pod "61729bed-368a-498d-b6f8-fcadf63835fc" (UID: "61729bed-368a-498d-b6f8-fcadf63835fc"). InnerVolumeSpecName "kube-api-access-kr8sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.472606 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "61729bed-368a-498d-b6f8-fcadf63835fc" (UID: "61729bed-368a-498d-b6f8-fcadf63835fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.540672 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.540714 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/61729bed-368a-498d-b6f8-fcadf63835fc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.540727 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr8sg\" (UniqueName: \"kubernetes.io/projected/61729bed-368a-498d-b6f8-fcadf63835fc-kube-api-access-kr8sg\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.933951 4708 generic.go:334] "Generic (PLEG): container finished" podID="61729bed-368a-498d-b6f8-fcadf63835fc" containerID="433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c" exitCode=0 Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.933998 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerDied","Data":"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c"} Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.934026 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-psxsl" event={"ID":"61729bed-368a-498d-b6f8-fcadf63835fc","Type":"ContainerDied","Data":"d0c3645f2b8c98188f72d5180976da69d8572d22dab00d789f868d3e0b075935"} Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.934043 4708 scope.go:117] "RemoveContainer" containerID="433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.934163 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-psxsl" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.955537 4708 scope.go:117] "RemoveContainer" containerID="2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.973933 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.987746 4708 scope.go:117] "RemoveContainer" containerID="362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb" Nov 25 06:21:17 crc kubenswrapper[4708]: I1125 06:21:17.991616 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-psxsl"] Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.014506 4708 scope.go:117] "RemoveContainer" containerID="433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c" Nov 25 06:21:18 crc kubenswrapper[4708]: E1125 06:21:18.015070 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c\": container with ID starting with 433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c not found: ID does not exist" containerID="433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.015161 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c"} err="failed to get container status \"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c\": rpc error: code = NotFound desc = could not find container \"433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c\": container with ID starting with 433e76c2d8f95e1639bde4c47da36a1d097e5ac237787e74ad8974e326fcd73c not found: ID does not exist" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.015244 4708 scope.go:117] "RemoveContainer" containerID="2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239" Nov 25 06:21:18 crc kubenswrapper[4708]: E1125 06:21:18.015662 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239\": container with ID starting with 2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239 not found: ID does not exist" containerID="2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.015694 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239"} err="failed to get container status \"2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239\": rpc error: code = NotFound desc = could not find container \"2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239\": container with ID starting with 2f9dca6665ae04f4c83eb74459c4a8eeea618747d99f30f62ab3dd0ff851a239 not found: ID does not exist" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.015716 4708 scope.go:117] "RemoveContainer" containerID="362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb" Nov 25 06:21:18 crc kubenswrapper[4708]: E1125 06:21:18.015964 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb\": container with ID starting with 362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb not found: ID does not exist" containerID="362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.015988 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb"} err="failed to get container status \"362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb\": rpc error: code = NotFound desc = could not find container \"362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb\": container with ID starting with 362cde6434acd44999bf5ba8a78daa4bc8bf9f11cea701b943b7c47d9742b3cb not found: ID does not exist" Nov 25 06:21:18 crc kubenswrapper[4708]: I1125 06:21:18.903221 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" path="/var/lib/kubelet/pods/61729bed-368a-498d-b6f8-fcadf63835fc/volumes" Nov 25 06:21:24 crc kubenswrapper[4708]: I1125 06:21:24.894501 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:21:24 crc kubenswrapper[4708]: E1125 06:21:24.895458 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.558313 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:30 crc kubenswrapper[4708]: E1125 06:21:30.560489 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="extract-content" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.560508 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="extract-content" Nov 25 06:21:30 crc kubenswrapper[4708]: E1125 06:21:30.560567 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="extract-utilities" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.560578 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="extract-utilities" Nov 25 06:21:30 crc kubenswrapper[4708]: E1125 06:21:30.560593 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="registry-server" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.560599 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="registry-server" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.560801 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="61729bed-368a-498d-b6f8-fcadf63835fc" containerName="registry-server" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.562225 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.571832 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.609936 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.610201 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.610392 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lh5l8\" (UniqueName: \"kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.712225 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.712302 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lh5l8\" (UniqueName: \"kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.712400 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.712888 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.712903 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.734220 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lh5l8\" (UniqueName: \"kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8\") pod \"redhat-marketplace-b6mlx\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:30 crc kubenswrapper[4708]: I1125 06:21:30.886180 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:31 crc kubenswrapper[4708]: I1125 06:21:31.313716 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:32 crc kubenswrapper[4708]: I1125 06:21:32.063114 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerID="d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b" exitCode=0 Nov 25 06:21:32 crc kubenswrapper[4708]: I1125 06:21:32.063184 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerDied","Data":"d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b"} Nov 25 06:21:32 crc kubenswrapper[4708]: I1125 06:21:32.063237 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerStarted","Data":"8f97a440fdc0ece8a8cb7f0694330a2d138b6b95361be01c4354fda7ad5a36e3"} Nov 25 06:21:33 crc kubenswrapper[4708]: I1125 06:21:33.077899 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerStarted","Data":"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4"} Nov 25 06:21:34 crc kubenswrapper[4708]: I1125 06:21:34.091876 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerID="67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4" exitCode=0 Nov 25 06:21:34 crc kubenswrapper[4708]: I1125 06:21:34.091985 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerDied","Data":"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4"} Nov 25 06:21:35 crc kubenswrapper[4708]: I1125 06:21:35.105137 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerStarted","Data":"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e"} Nov 25 06:21:35 crc kubenswrapper[4708]: I1125 06:21:35.132730 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b6mlx" podStartSLOduration=2.622884986 podStartE2EDuration="5.13269184s" podCreationTimestamp="2025-11-25 06:21:30 +0000 UTC" firstStartedPulling="2025-11-25 06:21:32.066105189 +0000 UTC m=+2433.474938575" lastFinishedPulling="2025-11-25 06:21:34.575912043 +0000 UTC m=+2435.984745429" observedRunningTime="2025-11-25 06:21:35.125574714 +0000 UTC m=+2436.534408100" watchObservedRunningTime="2025-11-25 06:21:35.13269184 +0000 UTC m=+2436.541525227" Nov 25 06:21:37 crc kubenswrapper[4708]: I1125 06:21:37.892975 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:21:37 crc kubenswrapper[4708]: E1125 06:21:37.894390 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:21:40 crc kubenswrapper[4708]: I1125 06:21:40.886632 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:40 crc kubenswrapper[4708]: I1125 06:21:40.886914 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:40 crc kubenswrapper[4708]: I1125 06:21:40.923774 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:41 crc kubenswrapper[4708]: I1125 06:21:41.186808 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:41 crc kubenswrapper[4708]: I1125 06:21:41.230478 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.167678 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b6mlx" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="registry-server" containerID="cri-o://8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e" gracePeriod=2 Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.553245 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.676314 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lh5l8\" (UniqueName: \"kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8\") pod \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.676485 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities\") pod \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.676707 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content\") pod \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\" (UID: \"a2c7856c-aa7a-41a0-9948-289fb3da98a9\") " Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.677258 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities" (OuterVolumeSpecName: "utilities") pod "a2c7856c-aa7a-41a0-9948-289fb3da98a9" (UID: "a2c7856c-aa7a-41a0-9948-289fb3da98a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.684595 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8" (OuterVolumeSpecName: "kube-api-access-lh5l8") pod "a2c7856c-aa7a-41a0-9948-289fb3da98a9" (UID: "a2c7856c-aa7a-41a0-9948-289fb3da98a9"). InnerVolumeSpecName "kube-api-access-lh5l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.690821 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2c7856c-aa7a-41a0-9948-289fb3da98a9" (UID: "a2c7856c-aa7a-41a0-9948-289fb3da98a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.778038 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.778069 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2c7856c-aa7a-41a0-9948-289fb3da98a9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:43 crc kubenswrapper[4708]: I1125 06:21:43.778081 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lh5l8\" (UniqueName: \"kubernetes.io/projected/a2c7856c-aa7a-41a0-9948-289fb3da98a9-kube-api-access-lh5l8\") on node \"crc\" DevicePath \"\"" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.175693 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerID="8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e" exitCode=0 Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.175737 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerDied","Data":"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e"} Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.175756 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b6mlx" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.175771 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b6mlx" event={"ID":"a2c7856c-aa7a-41a0-9948-289fb3da98a9","Type":"ContainerDied","Data":"8f97a440fdc0ece8a8cb7f0694330a2d138b6b95361be01c4354fda7ad5a36e3"} Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.175799 4708 scope.go:117] "RemoveContainer" containerID="8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.200980 4708 scope.go:117] "RemoveContainer" containerID="67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.217955 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.223811 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b6mlx"] Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.238113 4708 scope.go:117] "RemoveContainer" containerID="d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.257027 4708 scope.go:117] "RemoveContainer" containerID="8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e" Nov 25 06:21:44 crc kubenswrapper[4708]: E1125 06:21:44.257419 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e\": container with ID starting with 8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e not found: ID does not exist" containerID="8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.257473 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e"} err="failed to get container status \"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e\": rpc error: code = NotFound desc = could not find container \"8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e\": container with ID starting with 8bdececf9bd4b9fda5383a9310f293911877ab07a44bd30195970b411dc2de7e not found: ID does not exist" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.257498 4708 scope.go:117] "RemoveContainer" containerID="67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4" Nov 25 06:21:44 crc kubenswrapper[4708]: E1125 06:21:44.257799 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4\": container with ID starting with 67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4 not found: ID does not exist" containerID="67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.257837 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4"} err="failed to get container status \"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4\": rpc error: code = NotFound desc = could not find container \"67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4\": container with ID starting with 67cb586ba14553354fb9dcb7b2f14df9c69dc1594553651d356a3f229016f3a4 not found: ID does not exist" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.257872 4708 scope.go:117] "RemoveContainer" containerID="d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b" Nov 25 06:21:44 crc kubenswrapper[4708]: E1125 06:21:44.258120 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b\": container with ID starting with d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b not found: ID does not exist" containerID="d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.258150 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b"} err="failed to get container status \"d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b\": rpc error: code = NotFound desc = could not find container \"d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b\": container with ID starting with d7f3cd7915154ab2045595940d5557251aa705a41ce5e55a8457fc889ad7834b not found: ID does not exist" Nov 25 06:21:44 crc kubenswrapper[4708]: I1125 06:21:44.955170 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" path="/var/lib/kubelet/pods/a2c7856c-aa7a-41a0-9948-289fb3da98a9/volumes" Nov 25 06:21:50 crc kubenswrapper[4708]: I1125 06:21:50.893999 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:21:50 crc kubenswrapper[4708]: E1125 06:21:50.894712 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:22:04 crc kubenswrapper[4708]: I1125 06:22:04.894099 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:22:04 crc kubenswrapper[4708]: E1125 06:22:04.895444 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:22:15 crc kubenswrapper[4708]: I1125 06:22:15.893120 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:22:16 crc kubenswrapper[4708]: I1125 06:22:16.411915 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2"} Nov 25 06:24:44 crc kubenswrapper[4708]: I1125 06:24:44.356660 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:24:44 crc kubenswrapper[4708]: I1125 06:24:44.357081 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:25:14 crc kubenswrapper[4708]: I1125 06:25:14.356869 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:25:14 crc kubenswrapper[4708]: I1125 06:25:14.357567 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:25:44 crc kubenswrapper[4708]: I1125 06:25:44.357790 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:25:44 crc kubenswrapper[4708]: I1125 06:25:44.358645 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:25:44 crc kubenswrapper[4708]: I1125 06:25:44.358731 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:25:44 crc kubenswrapper[4708]: I1125 06:25:44.360125 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:25:44 crc kubenswrapper[4708]: I1125 06:25:44.360205 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2" gracePeriod=600 Nov 25 06:25:45 crc kubenswrapper[4708]: I1125 06:25:45.001964 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2" exitCode=0 Nov 25 06:25:45 crc kubenswrapper[4708]: I1125 06:25:45.002060 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2"} Nov 25 06:25:45 crc kubenswrapper[4708]: I1125 06:25:45.002350 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4"} Nov 25 06:25:45 crc kubenswrapper[4708]: I1125 06:25:45.002378 4708 scope.go:117] "RemoveContainer" containerID="d92348c9560a3cb5b92ad3c23b8a1396f131686fd0e6eaecc8e5c25b5d351197" Nov 25 06:27:44 crc kubenswrapper[4708]: I1125 06:27:44.357451 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:27:44 crc kubenswrapper[4708]: I1125 06:27:44.358273 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:28:14 crc kubenswrapper[4708]: I1125 06:28:14.357439 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:28:14 crc kubenswrapper[4708]: I1125 06:28:14.358029 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.745152 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:27 crc kubenswrapper[4708]: E1125 06:28:27.746115 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="extract-content" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.746130 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="extract-content" Nov 25 06:28:27 crc kubenswrapper[4708]: E1125 06:28:27.746158 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="registry-server" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.746164 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="registry-server" Nov 25 06:28:27 crc kubenswrapper[4708]: E1125 06:28:27.746191 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="extract-utilities" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.746199 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="extract-utilities" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.746392 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c7856c-aa7a-41a0-9948-289fb3da98a9" containerName="registry-server" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.747733 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.755558 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.949064 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz7jq\" (UniqueName: \"kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.949115 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:27 crc kubenswrapper[4708]: I1125 06:28:27.949409 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.051642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.052021 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz7jq\" (UniqueName: \"kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.052060 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.052640 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.052752 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.069685 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz7jq\" (UniqueName: \"kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq\") pod \"redhat-operators-l5fqc\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.077330 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:28 crc kubenswrapper[4708]: I1125 06:28:28.500904 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:29 crc kubenswrapper[4708]: I1125 06:28:29.407895 4708 generic.go:334] "Generic (PLEG): container finished" podID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerID="0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210" exitCode=0 Nov 25 06:28:29 crc kubenswrapper[4708]: I1125 06:28:29.408205 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerDied","Data":"0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210"} Nov 25 06:28:29 crc kubenswrapper[4708]: I1125 06:28:29.408291 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerStarted","Data":"6b471db20e7df62d4fd8df2f45ec04476aa06d2a8d4b9f15b5a7d68168a16e02"} Nov 25 06:28:29 crc kubenswrapper[4708]: I1125 06:28:29.410884 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:28:30 crc kubenswrapper[4708]: I1125 06:28:30.419805 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerStarted","Data":"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a"} Nov 25 06:28:31 crc kubenswrapper[4708]: I1125 06:28:31.434971 4708 generic.go:334] "Generic (PLEG): container finished" podID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerID="f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a" exitCode=0 Nov 25 06:28:31 crc kubenswrapper[4708]: I1125 06:28:31.435038 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerDied","Data":"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a"} Nov 25 06:28:32 crc kubenswrapper[4708]: I1125 06:28:32.446432 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerStarted","Data":"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780"} Nov 25 06:28:32 crc kubenswrapper[4708]: I1125 06:28:32.470849 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l5fqc" podStartSLOduration=2.911330263 podStartE2EDuration="5.470830156s" podCreationTimestamp="2025-11-25 06:28:27 +0000 UTC" firstStartedPulling="2025-11-25 06:28:29.410586431 +0000 UTC m=+2850.819419816" lastFinishedPulling="2025-11-25 06:28:31.970086322 +0000 UTC m=+2853.378919709" observedRunningTime="2025-11-25 06:28:32.460371269 +0000 UTC m=+2853.869204655" watchObservedRunningTime="2025-11-25 06:28:32.470830156 +0000 UTC m=+2853.879663542" Nov 25 06:28:38 crc kubenswrapper[4708]: I1125 06:28:38.077847 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:38 crc kubenswrapper[4708]: I1125 06:28:38.078619 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:38 crc kubenswrapper[4708]: I1125 06:28:38.120158 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:38 crc kubenswrapper[4708]: I1125 06:28:38.537192 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:38 crc kubenswrapper[4708]: I1125 06:28:38.595232 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.512703 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l5fqc" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="registry-server" containerID="cri-o://840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780" gracePeriod=2 Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.903547 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.930731 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz7jq\" (UniqueName: \"kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq\") pod \"8346837a-7daa-46d9-850c-3b5cb1958e7c\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.930876 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities\") pod \"8346837a-7daa-46d9-850c-3b5cb1958e7c\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.930933 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content\") pod \"8346837a-7daa-46d9-850c-3b5cb1958e7c\" (UID: \"8346837a-7daa-46d9-850c-3b5cb1958e7c\") " Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.931606 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities" (OuterVolumeSpecName: "utilities") pod "8346837a-7daa-46d9-850c-3b5cb1958e7c" (UID: "8346837a-7daa-46d9-850c-3b5cb1958e7c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.938336 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq" (OuterVolumeSpecName: "kube-api-access-bz7jq") pod "8346837a-7daa-46d9-850c-3b5cb1958e7c" (UID: "8346837a-7daa-46d9-850c-3b5cb1958e7c"). InnerVolumeSpecName "kube-api-access-bz7jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:28:40 crc kubenswrapper[4708]: I1125 06:28:40.999866 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8346837a-7daa-46d9-850c-3b5cb1958e7c" (UID: "8346837a-7daa-46d9-850c-3b5cb1958e7c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.032659 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz7jq\" (UniqueName: \"kubernetes.io/projected/8346837a-7daa-46d9-850c-3b5cb1958e7c-kube-api-access-bz7jq\") on node \"crc\" DevicePath \"\"" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.032689 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.032699 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8346837a-7daa-46d9-850c-3b5cb1958e7c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.522230 4708 generic.go:334] "Generic (PLEG): container finished" podID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerID="840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780" exitCode=0 Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.522310 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerDied","Data":"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780"} Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.522400 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l5fqc" event={"ID":"8346837a-7daa-46d9-850c-3b5cb1958e7c","Type":"ContainerDied","Data":"6b471db20e7df62d4fd8df2f45ec04476aa06d2a8d4b9f15b5a7d68168a16e02"} Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.522429 4708 scope.go:117] "RemoveContainer" containerID="840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.522330 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l5fqc" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.542140 4708 scope.go:117] "RemoveContainer" containerID="f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.559174 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.583679 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l5fqc"] Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.585951 4708 scope.go:117] "RemoveContainer" containerID="0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.606942 4708 scope.go:117] "RemoveContainer" containerID="840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780" Nov 25 06:28:41 crc kubenswrapper[4708]: E1125 06:28:41.607477 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780\": container with ID starting with 840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780 not found: ID does not exist" containerID="840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.607548 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780"} err="failed to get container status \"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780\": rpc error: code = NotFound desc = could not find container \"840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780\": container with ID starting with 840c7b271b0810b199dc2f43e5ed196cc1816059a21e3b47d52565b903d7b780 not found: ID does not exist" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.607578 4708 scope.go:117] "RemoveContainer" containerID="f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a" Nov 25 06:28:41 crc kubenswrapper[4708]: E1125 06:28:41.607894 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a\": container with ID starting with f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a not found: ID does not exist" containerID="f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.607929 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a"} err="failed to get container status \"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a\": rpc error: code = NotFound desc = could not find container \"f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a\": container with ID starting with f82f63c99bc62552b6bfce368d5cd187ccc77133bcb54bad02c723fb8d20705a not found: ID does not exist" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.607954 4708 scope.go:117] "RemoveContainer" containerID="0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210" Nov 25 06:28:41 crc kubenswrapper[4708]: E1125 06:28:41.608238 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210\": container with ID starting with 0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210 not found: ID does not exist" containerID="0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210" Nov 25 06:28:41 crc kubenswrapper[4708]: I1125 06:28:41.608270 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210"} err="failed to get container status \"0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210\": rpc error: code = NotFound desc = could not find container \"0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210\": container with ID starting with 0f0d382af704b9ba869da9d09e4c1a17d7dba8ffd5950571d38c1552bfdcf210 not found: ID does not exist" Nov 25 06:28:42 crc kubenswrapper[4708]: I1125 06:28:42.908368 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" path="/var/lib/kubelet/pods/8346837a-7daa-46d9-850c-3b5cb1958e7c/volumes" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.357264 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.357939 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.358024 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.359647 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.359726 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" gracePeriod=600 Nov 25 06:28:44 crc kubenswrapper[4708]: E1125 06:28:44.488773 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.566117 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" exitCode=0 Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.566174 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4"} Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.566218 4708 scope.go:117] "RemoveContainer" containerID="e7d4f0bd375afe5964753ca7ee599a0d1d40950f82ed0facec9396b829f6ebf2" Nov 25 06:28:44 crc kubenswrapper[4708]: I1125 06:28:44.567061 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:28:44 crc kubenswrapper[4708]: E1125 06:28:44.567325 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:28:57 crc kubenswrapper[4708]: I1125 06:28:57.892951 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:28:57 crc kubenswrapper[4708]: E1125 06:28:57.893686 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:29:11 crc kubenswrapper[4708]: I1125 06:29:11.894001 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:29:11 crc kubenswrapper[4708]: E1125 06:29:11.895108 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:29:24 crc kubenswrapper[4708]: I1125 06:29:24.893503 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:29:24 crc kubenswrapper[4708]: E1125 06:29:24.894416 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:29:37 crc kubenswrapper[4708]: I1125 06:29:37.893079 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:29:37 crc kubenswrapper[4708]: E1125 06:29:37.893861 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:29:52 crc kubenswrapper[4708]: I1125 06:29:52.893399 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:29:52 crc kubenswrapper[4708]: E1125 06:29:52.894681 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.140589 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt"] Nov 25 06:30:00 crc kubenswrapper[4708]: E1125 06:30:00.141515 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="extract-utilities" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.141540 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="extract-utilities" Nov 25 06:30:00 crc kubenswrapper[4708]: E1125 06:30:00.141548 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="registry-server" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.141555 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="registry-server" Nov 25 06:30:00 crc kubenswrapper[4708]: E1125 06:30:00.141586 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="extract-content" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.141592 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="extract-content" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.141768 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="8346837a-7daa-46d9-850c-3b5cb1958e7c" containerName="registry-server" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.142456 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.144861 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.145006 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.148975 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt"] Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.296431 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.296731 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.297071 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tn8f\" (UniqueName: \"kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.399713 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.399798 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tn8f\" (UniqueName: \"kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.399885 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.400900 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.407418 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.415881 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tn8f\" (UniqueName: \"kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f\") pod \"collect-profiles-29400870-2p4zt\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.467816 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:00 crc kubenswrapper[4708]: I1125 06:30:00.873794 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt"] Nov 25 06:30:01 crc kubenswrapper[4708]: I1125 06:30:01.264655 4708 generic.go:334] "Generic (PLEG): container finished" podID="3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" containerID="48691a3b43b8efd0aaec153a9577511fc19d0015ada8c6fb0a0ea66c164a1863" exitCode=0 Nov 25 06:30:01 crc kubenswrapper[4708]: I1125 06:30:01.264733 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" event={"ID":"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d","Type":"ContainerDied","Data":"48691a3b43b8efd0aaec153a9577511fc19d0015ada8c6fb0a0ea66c164a1863"} Nov 25 06:30:01 crc kubenswrapper[4708]: I1125 06:30:01.264796 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" event={"ID":"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d","Type":"ContainerStarted","Data":"a5e2edb3486af071ce5996d567cc1da3b69c2bdd2f76b5688f8dfba6f97a27d2"} Nov 25 06:30:01 crc kubenswrapper[4708]: E1125 06:30:01.348641 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f60d5e2_9732_4c37_ac5e_00f69c0beb2d.slice/crio-conmon-48691a3b43b8efd0aaec153a9577511fc19d0015ada8c6fb0a0ea66c164a1863.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3f60d5e2_9732_4c37_ac5e_00f69c0beb2d.slice/crio-48691a3b43b8efd0aaec153a9577511fc19d0015ada8c6fb0a0ea66c164a1863.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.534209 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.550609 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume\") pod \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.550726 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tn8f\" (UniqueName: \"kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f\") pod \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.550799 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume\") pod \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\" (UID: \"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d\") " Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.551312 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume" (OuterVolumeSpecName: "config-volume") pod "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" (UID: "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.556670 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f" (OuterVolumeSpecName: "kube-api-access-2tn8f") pod "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" (UID: "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d"). InnerVolumeSpecName "kube-api-access-2tn8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.556814 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" (UID: "3f60d5e2-9732-4c37-ac5e-00f69c0beb2d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.653556 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.653591 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tn8f\" (UniqueName: \"kubernetes.io/projected/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-kube-api-access-2tn8f\") on node \"crc\" DevicePath \"\"" Nov 25 06:30:02 crc kubenswrapper[4708]: I1125 06:30:02.653606 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f60d5e2-9732-4c37-ac5e-00f69c0beb2d-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.287788 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" event={"ID":"3f60d5e2-9732-4c37-ac5e-00f69c0beb2d","Type":"ContainerDied","Data":"a5e2edb3486af071ce5996d567cc1da3b69c2bdd2f76b5688f8dfba6f97a27d2"} Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.288120 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5e2edb3486af071ce5996d567cc1da3b69c2bdd2f76b5688f8dfba6f97a27d2" Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.287897 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400870-2p4zt" Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.608382 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2"] Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.613714 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400825-kljt2"] Nov 25 06:30:03 crc kubenswrapper[4708]: I1125 06:30:03.893921 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:30:03 crc kubenswrapper[4708]: E1125 06:30:03.894174 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:30:04 crc kubenswrapper[4708]: I1125 06:30:04.902867 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38ef401b-8325-4595-9d9e-dd6f0aca7078" path="/var/lib/kubelet/pods/38ef401b-8325-4595-9d9e-dd6f0aca7078/volumes" Nov 25 06:30:18 crc kubenswrapper[4708]: I1125 06:30:18.900342 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:30:18 crc kubenswrapper[4708]: E1125 06:30:18.901314 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:30:30 crc kubenswrapper[4708]: I1125 06:30:30.894372 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:30:30 crc kubenswrapper[4708]: E1125 06:30:30.895330 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:30:45 crc kubenswrapper[4708]: I1125 06:30:45.894351 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:30:45 crc kubenswrapper[4708]: E1125 06:30:45.895597 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:30:56 crc kubenswrapper[4708]: I1125 06:30:56.894278 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:30:56 crc kubenswrapper[4708]: E1125 06:30:56.895009 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:31:02 crc kubenswrapper[4708]: I1125 06:31:02.169903 4708 scope.go:117] "RemoveContainer" containerID="f6f52a6a86f8fcb5169e062b0da706dca690087810b1ceaeb1bd339f5b983e42" Nov 25 06:31:08 crc kubenswrapper[4708]: I1125 06:31:08.897992 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:31:08 crc kubenswrapper[4708]: E1125 06:31:08.898611 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:31:22 crc kubenswrapper[4708]: I1125 06:31:22.893685 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:31:22 crc kubenswrapper[4708]: E1125 06:31:22.894491 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.520761 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:25 crc kubenswrapper[4708]: E1125 06:31:25.522134 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" containerName="collect-profiles" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.522151 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" containerName="collect-profiles" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.522347 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f60d5e2-9732-4c37-ac5e-00f69c0beb2d" containerName="collect-profiles" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.528171 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.532506 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.579709 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.579765 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.579789 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvbqj\" (UniqueName: \"kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.681509 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.681594 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.681620 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvbqj\" (UniqueName: \"kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.681991 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.682041 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.698453 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvbqj\" (UniqueName: \"kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj\") pod \"certified-operators-4c5zx\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:25 crc kubenswrapper[4708]: I1125 06:31:25.854874 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:26 crc kubenswrapper[4708]: I1125 06:31:26.330958 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:27 crc kubenswrapper[4708]: I1125 06:31:27.009606 4708 generic.go:334] "Generic (PLEG): container finished" podID="25ad09cc-f060-41e3-a641-f7fee9798001" containerID="e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df" exitCode=0 Nov 25 06:31:27 crc kubenswrapper[4708]: I1125 06:31:27.009649 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerDied","Data":"e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df"} Nov 25 06:31:27 crc kubenswrapper[4708]: I1125 06:31:27.010114 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerStarted","Data":"f1e6de37416e027e3ecf7c078b0d328d0846e518f58f442bcbfe43476cea9423"} Nov 25 06:31:28 crc kubenswrapper[4708]: I1125 06:31:28.022901 4708 generic.go:334] "Generic (PLEG): container finished" podID="25ad09cc-f060-41e3-a641-f7fee9798001" containerID="2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14" exitCode=0 Nov 25 06:31:28 crc kubenswrapper[4708]: I1125 06:31:28.022951 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerDied","Data":"2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14"} Nov 25 06:31:29 crc kubenswrapper[4708]: I1125 06:31:29.033484 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerStarted","Data":"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e"} Nov 25 06:31:29 crc kubenswrapper[4708]: I1125 06:31:29.055312 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4c5zx" podStartSLOduration=2.433636132 podStartE2EDuration="4.055287543s" podCreationTimestamp="2025-11-25 06:31:25 +0000 UTC" firstStartedPulling="2025-11-25 06:31:27.011807507 +0000 UTC m=+3028.420640892" lastFinishedPulling="2025-11-25 06:31:28.633458917 +0000 UTC m=+3030.042292303" observedRunningTime="2025-11-25 06:31:29.050015326 +0000 UTC m=+3030.458848712" watchObservedRunningTime="2025-11-25 06:31:29.055287543 +0000 UTC m=+3030.464120930" Nov 25 06:31:35 crc kubenswrapper[4708]: I1125 06:31:35.856115 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:35 crc kubenswrapper[4708]: I1125 06:31:35.856874 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:35 crc kubenswrapper[4708]: I1125 06:31:35.896015 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:36 crc kubenswrapper[4708]: I1125 06:31:36.138512 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:36 crc kubenswrapper[4708]: I1125 06:31:36.185664 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:36 crc kubenswrapper[4708]: I1125 06:31:36.894139 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:31:36 crc kubenswrapper[4708]: E1125 06:31:36.896263 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.116611 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4c5zx" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="registry-server" containerID="cri-o://ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e" gracePeriod=2 Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.522716 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.554841 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities\") pod \"25ad09cc-f060-41e3-a641-f7fee9798001\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.554917 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content\") pod \"25ad09cc-f060-41e3-a641-f7fee9798001\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.554951 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvbqj\" (UniqueName: \"kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj\") pod \"25ad09cc-f060-41e3-a641-f7fee9798001\" (UID: \"25ad09cc-f060-41e3-a641-f7fee9798001\") " Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.555626 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities" (OuterVolumeSpecName: "utilities") pod "25ad09cc-f060-41e3-a641-f7fee9798001" (UID: "25ad09cc-f060-41e3-a641-f7fee9798001"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.563435 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj" (OuterVolumeSpecName: "kube-api-access-qvbqj") pod "25ad09cc-f060-41e3-a641-f7fee9798001" (UID: "25ad09cc-f060-41e3-a641-f7fee9798001"). InnerVolumeSpecName "kube-api-access-qvbqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.590232 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25ad09cc-f060-41e3-a641-f7fee9798001" (UID: "25ad09cc-f060-41e3-a641-f7fee9798001"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.656363 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.656392 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25ad09cc-f060-41e3-a641-f7fee9798001-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:31:38 crc kubenswrapper[4708]: I1125 06:31:38.656405 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvbqj\" (UniqueName: \"kubernetes.io/projected/25ad09cc-f060-41e3-a641-f7fee9798001-kube-api-access-qvbqj\") on node \"crc\" DevicePath \"\"" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.126375 4708 generic.go:334] "Generic (PLEG): container finished" podID="25ad09cc-f060-41e3-a641-f7fee9798001" containerID="ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e" exitCode=0 Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.126448 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4c5zx" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.126470 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerDied","Data":"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e"} Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.127169 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4c5zx" event={"ID":"25ad09cc-f060-41e3-a641-f7fee9798001","Type":"ContainerDied","Data":"f1e6de37416e027e3ecf7c078b0d328d0846e518f58f442bcbfe43476cea9423"} Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.127197 4708 scope.go:117] "RemoveContainer" containerID="ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.143268 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.145221 4708 scope.go:117] "RemoveContainer" containerID="2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.149980 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4c5zx"] Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.163224 4708 scope.go:117] "RemoveContainer" containerID="e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.199645 4708 scope.go:117] "RemoveContainer" containerID="ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e" Nov 25 06:31:39 crc kubenswrapper[4708]: E1125 06:31:39.200253 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e\": container with ID starting with ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e not found: ID does not exist" containerID="ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.200298 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e"} err="failed to get container status \"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e\": rpc error: code = NotFound desc = could not find container \"ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e\": container with ID starting with ef4afd91642137a8538acb6dc39b79d983f09121ab8ce8fb83ae7f48dd9e158e not found: ID does not exist" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.200327 4708 scope.go:117] "RemoveContainer" containerID="2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14" Nov 25 06:31:39 crc kubenswrapper[4708]: E1125 06:31:39.200756 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14\": container with ID starting with 2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14 not found: ID does not exist" containerID="2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.200784 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14"} err="failed to get container status \"2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14\": rpc error: code = NotFound desc = could not find container \"2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14\": container with ID starting with 2743585f76f14b467c272ac33f78514b7b63465f1cfe1d97f039d4e5ebaa6a14 not found: ID does not exist" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.200801 4708 scope.go:117] "RemoveContainer" containerID="e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df" Nov 25 06:31:39 crc kubenswrapper[4708]: E1125 06:31:39.201152 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df\": container with ID starting with e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df not found: ID does not exist" containerID="e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df" Nov 25 06:31:39 crc kubenswrapper[4708]: I1125 06:31:39.201174 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df"} err="failed to get container status \"e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df\": rpc error: code = NotFound desc = could not find container \"e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df\": container with ID starting with e1eae7c7770c5e7ab8c5ecad3c2868ef3371c1e69c0cf4cfc317973441f991df not found: ID does not exist" Nov 25 06:31:40 crc kubenswrapper[4708]: I1125 06:31:40.903800 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" path="/var/lib/kubelet/pods/25ad09cc-f060-41e3-a641-f7fee9798001/volumes" Nov 25 06:31:47 crc kubenswrapper[4708]: I1125 06:31:47.893731 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:31:47 crc kubenswrapper[4708]: E1125 06:31:47.894972 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:31:59 crc kubenswrapper[4708]: I1125 06:31:59.894356 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:31:59 crc kubenswrapper[4708]: E1125 06:31:59.895346 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.841567 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:07 crc kubenswrapper[4708]: E1125 06:32:07.842447 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="registry-server" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.842461 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="registry-server" Nov 25 06:32:07 crc kubenswrapper[4708]: E1125 06:32:07.842503 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="extract-content" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.842509 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="extract-content" Nov 25 06:32:07 crc kubenswrapper[4708]: E1125 06:32:07.842542 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="extract-utilities" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.842577 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="extract-utilities" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.842744 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="25ad09cc-f060-41e3-a641-f7fee9798001" containerName="registry-server" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.843987 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.855365 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.965729 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.965803 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v256l\" (UniqueName: \"kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:07 crc kubenswrapper[4708]: I1125 06:32:07.965930 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.068702 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.069319 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.069642 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.070002 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.070081 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v256l\" (UniqueName: \"kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.092487 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v256l\" (UniqueName: \"kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l\") pod \"redhat-marketplace-4dsdh\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.162892 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:08 crc kubenswrapper[4708]: I1125 06:32:08.579332 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:09 crc kubenswrapper[4708]: I1125 06:32:09.396994 4708 generic.go:334] "Generic (PLEG): container finished" podID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerID="6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1" exitCode=0 Nov 25 06:32:09 crc kubenswrapper[4708]: I1125 06:32:09.397170 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerDied","Data":"6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1"} Nov 25 06:32:09 crc kubenswrapper[4708]: I1125 06:32:09.397461 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerStarted","Data":"12a08902d14e5a6dcf09defa34be5f0684f211c75f9162c6aae2631677f9ca28"} Nov 25 06:32:10 crc kubenswrapper[4708]: I1125 06:32:10.409060 4708 generic.go:334] "Generic (PLEG): container finished" podID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerID="b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b" exitCode=0 Nov 25 06:32:10 crc kubenswrapper[4708]: I1125 06:32:10.409277 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerDied","Data":"b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b"} Nov 25 06:32:11 crc kubenswrapper[4708]: I1125 06:32:11.422679 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerStarted","Data":"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04"} Nov 25 06:32:11 crc kubenswrapper[4708]: I1125 06:32:11.437544 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4dsdh" podStartSLOduration=2.897119891 podStartE2EDuration="4.437503069s" podCreationTimestamp="2025-11-25 06:32:07 +0000 UTC" firstStartedPulling="2025-11-25 06:32:09.399690926 +0000 UTC m=+3070.808524312" lastFinishedPulling="2025-11-25 06:32:10.940074104 +0000 UTC m=+3072.348907490" observedRunningTime="2025-11-25 06:32:11.437092145 +0000 UTC m=+3072.845925531" watchObservedRunningTime="2025-11-25 06:32:11.437503069 +0000 UTC m=+3072.846336456" Nov 25 06:32:14 crc kubenswrapper[4708]: I1125 06:32:14.894652 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:32:14 crc kubenswrapper[4708]: E1125 06:32:14.895951 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:32:18 crc kubenswrapper[4708]: I1125 06:32:18.163023 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:18 crc kubenswrapper[4708]: I1125 06:32:18.163427 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:18 crc kubenswrapper[4708]: I1125 06:32:18.200348 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:18 crc kubenswrapper[4708]: I1125 06:32:18.538103 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:18 crc kubenswrapper[4708]: I1125 06:32:18.585572 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:20 crc kubenswrapper[4708]: I1125 06:32:20.517612 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4dsdh" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="registry-server" containerID="cri-o://c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04" gracePeriod=2 Nov 25 06:32:20 crc kubenswrapper[4708]: I1125 06:32:20.917787 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.036256 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content\") pod \"538b5a7e-0bac-497c-bac3-2551a59b7d81\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.036493 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v256l\" (UniqueName: \"kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l\") pod \"538b5a7e-0bac-497c-bac3-2551a59b7d81\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.037249 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities\") pod \"538b5a7e-0bac-497c-bac3-2551a59b7d81\" (UID: \"538b5a7e-0bac-497c-bac3-2551a59b7d81\") " Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.038257 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities" (OuterVolumeSpecName: "utilities") pod "538b5a7e-0bac-497c-bac3-2551a59b7d81" (UID: "538b5a7e-0bac-497c-bac3-2551a59b7d81"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.042488 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l" (OuterVolumeSpecName: "kube-api-access-v256l") pod "538b5a7e-0bac-497c-bac3-2551a59b7d81" (UID: "538b5a7e-0bac-497c-bac3-2551a59b7d81"). InnerVolumeSpecName "kube-api-access-v256l". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.052646 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "538b5a7e-0bac-497c-bac3-2551a59b7d81" (UID: "538b5a7e-0bac-497c-bac3-2551a59b7d81"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.138724 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v256l\" (UniqueName: \"kubernetes.io/projected/538b5a7e-0bac-497c-bac3-2551a59b7d81-kube-api-access-v256l\") on node \"crc\" DevicePath \"\"" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.138945 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.138954 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/538b5a7e-0bac-497c-bac3-2551a59b7d81-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.531217 4708 generic.go:334] "Generic (PLEG): container finished" podID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerID="c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04" exitCode=0 Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.531275 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerDied","Data":"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04"} Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.531316 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4dsdh" event={"ID":"538b5a7e-0bac-497c-bac3-2551a59b7d81","Type":"ContainerDied","Data":"12a08902d14e5a6dcf09defa34be5f0684f211c75f9162c6aae2631677f9ca28"} Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.531339 4708 scope.go:117] "RemoveContainer" containerID="c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.531432 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4dsdh" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.555740 4708 scope.go:117] "RemoveContainer" containerID="b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.569243 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.581637 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4dsdh"] Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.588733 4708 scope.go:117] "RemoveContainer" containerID="6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.615491 4708 scope.go:117] "RemoveContainer" containerID="c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04" Nov 25 06:32:21 crc kubenswrapper[4708]: E1125 06:32:21.616028 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04\": container with ID starting with c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04 not found: ID does not exist" containerID="c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.616072 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04"} err="failed to get container status \"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04\": rpc error: code = NotFound desc = could not find container \"c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04\": container with ID starting with c75aad542a83d42095788a7710fd9745f8ab51e78969aecc64d400ea40bddd04 not found: ID does not exist" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.616099 4708 scope.go:117] "RemoveContainer" containerID="b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b" Nov 25 06:32:21 crc kubenswrapper[4708]: E1125 06:32:21.616510 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b\": container with ID starting with b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b not found: ID does not exist" containerID="b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.616568 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b"} err="failed to get container status \"b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b\": rpc error: code = NotFound desc = could not find container \"b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b\": container with ID starting with b81a6b42b1e59696f4690a71342a888b80eb3842589ab4c1df5dfb2203789a5b not found: ID does not exist" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.616604 4708 scope.go:117] "RemoveContainer" containerID="6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1" Nov 25 06:32:21 crc kubenswrapper[4708]: E1125 06:32:21.616925 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1\": container with ID starting with 6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1 not found: ID does not exist" containerID="6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1" Nov 25 06:32:21 crc kubenswrapper[4708]: I1125 06:32:21.616956 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1"} err="failed to get container status \"6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1\": rpc error: code = NotFound desc = could not find container \"6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1\": container with ID starting with 6bab8a3817cf448ac831edad2f1bdac880e670df9ae203fe82b40cfe91e7b7f1 not found: ID does not exist" Nov 25 06:32:22 crc kubenswrapper[4708]: I1125 06:32:22.904909 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" path="/var/lib/kubelet/pods/538b5a7e-0bac-497c-bac3-2551a59b7d81/volumes" Nov 25 06:32:29 crc kubenswrapper[4708]: I1125 06:32:29.893857 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:32:29 crc kubenswrapper[4708]: E1125 06:32:29.894846 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:32:40 crc kubenswrapper[4708]: I1125 06:32:40.893448 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:32:40 crc kubenswrapper[4708]: E1125 06:32:40.894434 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:32:54 crc kubenswrapper[4708]: I1125 06:32:54.892755 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:32:54 crc kubenswrapper[4708]: E1125 06:32:54.893608 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:33:07 crc kubenswrapper[4708]: I1125 06:33:07.893573 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:33:07 crc kubenswrapper[4708]: E1125 06:33:07.894623 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:33:18 crc kubenswrapper[4708]: I1125 06:33:18.900096 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:33:18 crc kubenswrapper[4708]: E1125 06:33:18.901166 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:33:31 crc kubenswrapper[4708]: I1125 06:33:31.893459 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:33:31 crc kubenswrapper[4708]: E1125 06:33:31.894213 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:33:43 crc kubenswrapper[4708]: I1125 06:33:43.893646 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:33:43 crc kubenswrapper[4708]: E1125 06:33:43.894632 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:33:54 crc kubenswrapper[4708]: I1125 06:33:54.893696 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:33:55 crc kubenswrapper[4708]: I1125 06:33:55.452740 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30"} Nov 25 06:34:12 crc kubenswrapper[4708]: I1125 06:34:12.608774 4708 generic.go:334] "Generic (PLEG): container finished" podID="695e5c80-15cf-4dc5-b286-cb10e12e9fee" containerID="9b53254b7e4efb4305fe6599eb57396df1ac06885e702b8adeaef13814ed51de" exitCode=0 Nov 25 06:34:12 crc kubenswrapper[4708]: I1125 06:34:12.608842 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"695e5c80-15cf-4dc5-b286-cb10e12e9fee","Type":"ContainerDied","Data":"9b53254b7e4efb4305fe6599eb57396df1ac06885e702b8adeaef13814ed51de"} Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.882257 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.921660 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.921895 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.921991 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.922094 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vcc59\" (UniqueName: \"kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.922215 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.922294 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.922381 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.923500 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data" (OuterVolumeSpecName: "config-data") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.923792 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.927832 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.928473 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.928887 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key\") pod \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\" (UID: \"695e5c80-15cf-4dc5-b286-cb10e12e9fee\") " Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.929632 4708 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.929775 4708 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.929861 4708 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/695e5c80-15cf-4dc5-b286-cb10e12e9fee-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.929783 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.931095 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59" (OuterVolumeSpecName: "kube-api-access-vcc59") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "kube-api-access-vcc59". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.951866 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.952703 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.954537 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:34:13 crc kubenswrapper[4708]: I1125 06:34:13.979246 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "695e5c80-15cf-4dc5-b286-cb10e12e9fee" (UID: "695e5c80-15cf-4dc5-b286-cb10e12e9fee"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032187 4708 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032226 4708 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ca-certs\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032237 4708 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032248 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032260 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vcc59\" (UniqueName: \"kubernetes.io/projected/695e5c80-15cf-4dc5-b286-cb10e12e9fee-kube-api-access-vcc59\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.032270 4708 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/695e5c80-15cf-4dc5-b286-cb10e12e9fee-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.049270 4708 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.134321 4708 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.626375 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"695e5c80-15cf-4dc5-b286-cb10e12e9fee","Type":"ContainerDied","Data":"0e396f85a4bd65f1b4307ef2aba254d2728c79adb2b3b842d3394b51f87e2428"} Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.626422 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e396f85a4bd65f1b4307ef2aba254d2728c79adb2b3b842d3394b51f87e2428" Nov 25 06:34:14 crc kubenswrapper[4708]: I1125 06:34:14.626457 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.738207 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 06:34:17 crc kubenswrapper[4708]: E1125 06:34:17.739141 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="registry-server" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739155 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="registry-server" Nov 25 06:34:17 crc kubenswrapper[4708]: E1125 06:34:17.739181 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="extract-content" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739186 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="extract-content" Nov 25 06:34:17 crc kubenswrapper[4708]: E1125 06:34:17.739199 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="extract-utilities" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739206 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="extract-utilities" Nov 25 06:34:17 crc kubenswrapper[4708]: E1125 06:34:17.739222 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="695e5c80-15cf-4dc5-b286-cb10e12e9fee" containerName="tempest-tests-tempest-tests-runner" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739226 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="695e5c80-15cf-4dc5-b286-cb10e12e9fee" containerName="tempest-tests-tempest-tests-runner" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739392 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="695e5c80-15cf-4dc5-b286-cb10e12e9fee" containerName="tempest-tests-tempest-tests-runner" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.739403 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="538b5a7e-0bac-497c-bac3-2551a59b7d81" containerName="registry-server" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.740844 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.747002 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-5cn88" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.749556 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.893738 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.894207 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svcjs\" (UniqueName: \"kubernetes.io/projected/655d4e8f-367a-475f-8363-1506473bb0d4-kube-api-access-svcjs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.997661 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svcjs\" (UniqueName: \"kubernetes.io/projected/655d4e8f-367a-475f-8363-1506473bb0d4-kube-api-access-svcjs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.997829 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:17 crc kubenswrapper[4708]: I1125 06:34:17.998351 4708 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.017901 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svcjs\" (UniqueName: \"kubernetes.io/projected/655d4e8f-367a-475f-8363-1506473bb0d4-kube-api-access-svcjs\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.022368 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"655d4e8f-367a-475f-8363-1506473bb0d4\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.067726 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.472280 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.478142 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:34:18 crc kubenswrapper[4708]: I1125 06:34:18.662415 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"655d4e8f-367a-475f-8363-1506473bb0d4","Type":"ContainerStarted","Data":"437cc544f5a78603f48aa7df0df84dbe5db140d20652cc868d37e00ec85e3dfe"} Nov 25 06:34:19 crc kubenswrapper[4708]: I1125 06:34:19.674500 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"655d4e8f-367a-475f-8363-1506473bb0d4","Type":"ContainerStarted","Data":"0f8243a293e33041b3f0d0e6dc2373b00378bcad89241621c70a7bd03a7aa2a3"} Nov 25 06:34:19 crc kubenswrapper[4708]: I1125 06:34:19.690879 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.8713611989999999 podStartE2EDuration="2.690854454s" podCreationTimestamp="2025-11-25 06:34:17 +0000 UTC" firstStartedPulling="2025-11-25 06:34:18.477851846 +0000 UTC m=+3199.886685231" lastFinishedPulling="2025-11-25 06:34:19.2973451 +0000 UTC m=+3200.706178486" observedRunningTime="2025-11-25 06:34:19.687672787 +0000 UTC m=+3201.096506172" watchObservedRunningTime="2025-11-25 06:34:19.690854454 +0000 UTC m=+3201.099687840" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.199639 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.202287 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.206911 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.337507 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.337899 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgbjq\" (UniqueName: \"kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.338217 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.440129 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgbjq\" (UniqueName: \"kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.440243 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.440300 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.440937 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.441080 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.477344 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgbjq\" (UniqueName: \"kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq\") pod \"community-operators-7d8d4\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.517451 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:30 crc kubenswrapper[4708]: I1125 06:34:30.975281 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:31 crc kubenswrapper[4708]: I1125 06:34:31.791888 4708 generic.go:334] "Generic (PLEG): container finished" podID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerID="4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc" exitCode=0 Nov 25 06:34:31 crc kubenswrapper[4708]: I1125 06:34:31.791950 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerDied","Data":"4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc"} Nov 25 06:34:31 crc kubenswrapper[4708]: I1125 06:34:31.792228 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerStarted","Data":"653c718fb4c7f2340f4a1af3bd3d48a57b86fce9d4da7c4aa69c8165ad2d4d4a"} Nov 25 06:34:32 crc kubenswrapper[4708]: I1125 06:34:32.806859 4708 generic.go:334] "Generic (PLEG): container finished" podID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerID="7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf" exitCode=0 Nov 25 06:34:32 crc kubenswrapper[4708]: I1125 06:34:32.806933 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerDied","Data":"7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf"} Nov 25 06:34:33 crc kubenswrapper[4708]: I1125 06:34:33.820949 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerStarted","Data":"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62"} Nov 25 06:34:33 crc kubenswrapper[4708]: I1125 06:34:33.838592 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7d8d4" podStartSLOduration=2.337681285 podStartE2EDuration="3.838564144s" podCreationTimestamp="2025-11-25 06:34:30 +0000 UTC" firstStartedPulling="2025-11-25 06:34:31.794867909 +0000 UTC m=+3213.203701296" lastFinishedPulling="2025-11-25 06:34:33.295750769 +0000 UTC m=+3214.704584155" observedRunningTime="2025-11-25 06:34:33.835938504 +0000 UTC m=+3215.244771891" watchObservedRunningTime="2025-11-25 06:34:33.838564144 +0000 UTC m=+3215.247397530" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.428185 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vgqw7/must-gather-hgfkx"] Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.430350 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.432339 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vgqw7"/"default-dockercfg-zcbrm" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.432597 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vgqw7"/"kube-root-ca.crt" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.433145 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vgqw7"/"openshift-service-ca.crt" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.456061 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vgqw7/must-gather-hgfkx"] Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.510754 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.511116 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p4gn\" (UniqueName: \"kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.613091 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.613158 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p4gn\" (UniqueName: \"kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.613767 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.645257 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p4gn\" (UniqueName: \"kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn\") pod \"must-gather-hgfkx\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:38 crc kubenswrapper[4708]: I1125 06:34:38.756452 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:34:39 crc kubenswrapper[4708]: I1125 06:34:39.185801 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vgqw7/must-gather-hgfkx"] Nov 25 06:34:39 crc kubenswrapper[4708]: W1125 06:34:39.188003 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod574fa4f1_02e8_4576_ac43_7a97950d5431.slice/crio-53c4c12686d969099c9afc0ac5edd2d72165cfb253d3605e4b44b1ee792060ea WatchSource:0}: Error finding container 53c4c12686d969099c9afc0ac5edd2d72165cfb253d3605e4b44b1ee792060ea: Status 404 returned error can't find the container with id 53c4c12686d969099c9afc0ac5edd2d72165cfb253d3605e4b44b1ee792060ea Nov 25 06:34:39 crc kubenswrapper[4708]: I1125 06:34:39.877187 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" event={"ID":"574fa4f1-02e8-4576-ac43-7a97950d5431","Type":"ContainerStarted","Data":"53c4c12686d969099c9afc0ac5edd2d72165cfb253d3605e4b44b1ee792060ea"} Nov 25 06:34:40 crc kubenswrapper[4708]: I1125 06:34:40.517773 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:40 crc kubenswrapper[4708]: I1125 06:34:40.517855 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:40 crc kubenswrapper[4708]: I1125 06:34:40.580976 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:40 crc kubenswrapper[4708]: I1125 06:34:40.936688 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:41 crc kubenswrapper[4708]: I1125 06:34:41.187598 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:42 crc kubenswrapper[4708]: I1125 06:34:42.904223 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7d8d4" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="registry-server" containerID="cri-o://37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62" gracePeriod=2 Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.399064 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.424640 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgbjq\" (UniqueName: \"kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq\") pod \"d80b46b0-85af-428d-a7ac-94f4a1019561\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.424817 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content\") pod \"d80b46b0-85af-428d-a7ac-94f4a1019561\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.424978 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities\") pod \"d80b46b0-85af-428d-a7ac-94f4a1019561\" (UID: \"d80b46b0-85af-428d-a7ac-94f4a1019561\") " Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.425614 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities" (OuterVolumeSpecName: "utilities") pod "d80b46b0-85af-428d-a7ac-94f4a1019561" (UID: "d80b46b0-85af-428d-a7ac-94f4a1019561"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.426392 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.433756 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq" (OuterVolumeSpecName: "kube-api-access-pgbjq") pod "d80b46b0-85af-428d-a7ac-94f4a1019561" (UID: "d80b46b0-85af-428d-a7ac-94f4a1019561"). InnerVolumeSpecName "kube-api-access-pgbjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.470034 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d80b46b0-85af-428d-a7ac-94f4a1019561" (UID: "d80b46b0-85af-428d-a7ac-94f4a1019561"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.528820 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgbjq\" (UniqueName: \"kubernetes.io/projected/d80b46b0-85af-428d-a7ac-94f4a1019561-kube-api-access-pgbjq\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.529076 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d80b46b0-85af-428d-a7ac-94f4a1019561-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.913816 4708 generic.go:334] "Generic (PLEG): container finished" podID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerID="37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62" exitCode=0 Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.913856 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerDied","Data":"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62"} Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.913883 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7d8d4" event={"ID":"d80b46b0-85af-428d-a7ac-94f4a1019561","Type":"ContainerDied","Data":"653c718fb4c7f2340f4a1af3bd3d48a57b86fce9d4da7c4aa69c8165ad2d4d4a"} Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.913904 4708 scope.go:117] "RemoveContainer" containerID="37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.914019 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7d8d4" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.943654 4708 scope.go:117] "RemoveContainer" containerID="7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf" Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.945617 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.951527 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7d8d4"] Nov 25 06:34:43 crc kubenswrapper[4708]: I1125 06:34:43.986031 4708 scope.go:117] "RemoveContainer" containerID="4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.004228 4708 scope.go:117] "RemoveContainer" containerID="37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62" Nov 25 06:34:44 crc kubenswrapper[4708]: E1125 06:34:44.004645 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62\": container with ID starting with 37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62 not found: ID does not exist" containerID="37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.004683 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62"} err="failed to get container status \"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62\": rpc error: code = NotFound desc = could not find container \"37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62\": container with ID starting with 37b4f075ebc25c39bba48e78f740568b37297531be22a42898235a897f33ca62 not found: ID does not exist" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.004708 4708 scope.go:117] "RemoveContainer" containerID="7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf" Nov 25 06:34:44 crc kubenswrapper[4708]: E1125 06:34:44.005105 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf\": container with ID starting with 7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf not found: ID does not exist" containerID="7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.005133 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf"} err="failed to get container status \"7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf\": rpc error: code = NotFound desc = could not find container \"7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf\": container with ID starting with 7695c5042302be1a7457c6e3ebf60391b461b8a7d523490b99627e0d801e2daf not found: ID does not exist" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.005151 4708 scope.go:117] "RemoveContainer" containerID="4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc" Nov 25 06:34:44 crc kubenswrapper[4708]: E1125 06:34:44.005402 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc\": container with ID starting with 4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc not found: ID does not exist" containerID="4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.005424 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc"} err="failed to get container status \"4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc\": rpc error: code = NotFound desc = could not find container \"4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc\": container with ID starting with 4cf59a3046e86d2041222f83d9c4f34d61c2a8e8d125fc0bf108305bbf64c8fc not found: ID does not exist" Nov 25 06:34:44 crc kubenswrapper[4708]: I1125 06:34:44.912089 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" path="/var/lib/kubelet/pods/d80b46b0-85af-428d-a7ac-94f4a1019561/volumes" Nov 25 06:34:45 crc kubenswrapper[4708]: I1125 06:34:45.942041 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" event={"ID":"574fa4f1-02e8-4576-ac43-7a97950d5431","Type":"ContainerStarted","Data":"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766"} Nov 25 06:34:46 crc kubenswrapper[4708]: I1125 06:34:46.953699 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" event={"ID":"574fa4f1-02e8-4576-ac43-7a97950d5431","Type":"ContainerStarted","Data":"43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821"} Nov 25 06:34:46 crc kubenswrapper[4708]: I1125 06:34:46.978347 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" podStartSLOduration=2.465075533 podStartE2EDuration="8.978326621s" podCreationTimestamp="2025-11-25 06:34:38 +0000 UTC" firstStartedPulling="2025-11-25 06:34:39.190615746 +0000 UTC m=+3220.599449132" lastFinishedPulling="2025-11-25 06:34:45.703866824 +0000 UTC m=+3227.112700220" observedRunningTime="2025-11-25 06:34:46.967961913 +0000 UTC m=+3228.376795299" watchObservedRunningTime="2025-11-25 06:34:46.978326621 +0000 UTC m=+3228.387159997" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.232403 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-59dsw"] Nov 25 06:34:49 crc kubenswrapper[4708]: E1125 06:34:49.233046 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="registry-server" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.233060 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="registry-server" Nov 25 06:34:49 crc kubenswrapper[4708]: E1125 06:34:49.233081 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="extract-content" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.233087 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="extract-content" Nov 25 06:34:49 crc kubenswrapper[4708]: E1125 06:34:49.233106 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="extract-utilities" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.233112 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="extract-utilities" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.233288 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="d80b46b0-85af-428d-a7ac-94f4a1019561" containerName="registry-server" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.233902 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.369426 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpvdw\" (UniqueName: \"kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.369943 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.471596 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpvdw\" (UniqueName: \"kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.471758 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.471831 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.488871 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpvdw\" (UniqueName: \"kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw\") pod \"crc-debug-59dsw\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.552425 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:34:49 crc kubenswrapper[4708]: W1125 06:34:49.586990 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82e6626e_4fda_4bf9_86a1_505c8535a6c4.slice/crio-9054615487c81453627ba031242d7ec7945ee7fa5cb0e63407504eb3542af1a5 WatchSource:0}: Error finding container 9054615487c81453627ba031242d7ec7945ee7fa5cb0e63407504eb3542af1a5: Status 404 returned error can't find the container with id 9054615487c81453627ba031242d7ec7945ee7fa5cb0e63407504eb3542af1a5 Nov 25 06:34:49 crc kubenswrapper[4708]: I1125 06:34:49.985575 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" event={"ID":"82e6626e-4fda-4bf9-86a1-505c8535a6c4","Type":"ContainerStarted","Data":"9054615487c81453627ba031242d7ec7945ee7fa5cb0e63407504eb3542af1a5"} Nov 25 06:35:03 crc kubenswrapper[4708]: I1125 06:35:03.099218 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" event={"ID":"82e6626e-4fda-4bf9-86a1-505c8535a6c4","Type":"ContainerStarted","Data":"731b10a3e32d4f5f25d2c4a241a9de57dd94494e2427f32f2e7756e3aa445dc6"} Nov 25 06:35:03 crc kubenswrapper[4708]: I1125 06:35:03.119633 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" podStartSLOduration=1.260612596 podStartE2EDuration="14.119615974s" podCreationTimestamp="2025-11-25 06:34:49 +0000 UTC" firstStartedPulling="2025-11-25 06:34:49.5889673 +0000 UTC m=+3230.997800685" lastFinishedPulling="2025-11-25 06:35:02.447970677 +0000 UTC m=+3243.856804063" observedRunningTime="2025-11-25 06:35:03.113426347 +0000 UTC m=+3244.522259733" watchObservedRunningTime="2025-11-25 06:35:03.119615974 +0000 UTC m=+3244.528449360" Nov 25 06:35:20 crc kubenswrapper[4708]: I1125 06:35:20.253335 4708 generic.go:334] "Generic (PLEG): container finished" podID="82e6626e-4fda-4bf9-86a1-505c8535a6c4" containerID="731b10a3e32d4f5f25d2c4a241a9de57dd94494e2427f32f2e7756e3aa445dc6" exitCode=0 Nov 25 06:35:20 crc kubenswrapper[4708]: I1125 06:35:20.253420 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" event={"ID":"82e6626e-4fda-4bf9-86a1-505c8535a6c4","Type":"ContainerDied","Data":"731b10a3e32d4f5f25d2c4a241a9de57dd94494e2427f32f2e7756e3aa445dc6"} Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.350018 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.375468 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-59dsw"] Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.380713 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-59dsw"] Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.472870 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpvdw\" (UniqueName: \"kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw\") pod \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.473426 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host\") pod \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\" (UID: \"82e6626e-4fda-4bf9-86a1-505c8535a6c4\") " Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.473498 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host" (OuterVolumeSpecName: "host") pod "82e6626e-4fda-4bf9-86a1-505c8535a6c4" (UID: "82e6626e-4fda-4bf9-86a1-505c8535a6c4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.473879 4708 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/82e6626e-4fda-4bf9-86a1-505c8535a6c4-host\") on node \"crc\" DevicePath \"\"" Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.480857 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw" (OuterVolumeSpecName: "kube-api-access-fpvdw") pod "82e6626e-4fda-4bf9-86a1-505c8535a6c4" (UID: "82e6626e-4fda-4bf9-86a1-505c8535a6c4"). InnerVolumeSpecName "kube-api-access-fpvdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:35:21 crc kubenswrapper[4708]: I1125 06:35:21.576006 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpvdw\" (UniqueName: \"kubernetes.io/projected/82e6626e-4fda-4bf9-86a1-505c8535a6c4-kube-api-access-fpvdw\") on node \"crc\" DevicePath \"\"" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.274369 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9054615487c81453627ba031242d7ec7945ee7fa5cb0e63407504eb3542af1a5" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.274434 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-59dsw" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.562225 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-zqj29"] Nov 25 06:35:22 crc kubenswrapper[4708]: E1125 06:35:22.562671 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e6626e-4fda-4bf9-86a1-505c8535a6c4" containerName="container-00" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.562686 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e6626e-4fda-4bf9-86a1-505c8535a6c4" containerName="container-00" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.562902 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e6626e-4fda-4bf9-86a1-505c8535a6c4" containerName="container-00" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.563572 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.696278 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7fvp\" (UniqueName: \"kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.696328 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.798269 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7fvp\" (UniqueName: \"kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.798320 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.798459 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.813072 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7fvp\" (UniqueName: \"kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp\") pod \"crc-debug-zqj29\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.878282 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:22 crc kubenswrapper[4708]: W1125 06:35:22.906206 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb26c5e08_b3b5_4f4a_9291_53a21b31f547.slice/crio-88242eea158383e3bb56f76c7a8d970309b658e02901e057f4c02ebbe4404ff7 WatchSource:0}: Error finding container 88242eea158383e3bb56f76c7a8d970309b658e02901e057f4c02ebbe4404ff7: Status 404 returned error can't find the container with id 88242eea158383e3bb56f76c7a8d970309b658e02901e057f4c02ebbe4404ff7 Nov 25 06:35:22 crc kubenswrapper[4708]: I1125 06:35:22.909996 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82e6626e-4fda-4bf9-86a1-505c8535a6c4" path="/var/lib/kubelet/pods/82e6626e-4fda-4bf9-86a1-505c8535a6c4/volumes" Nov 25 06:35:23 crc kubenswrapper[4708]: I1125 06:35:23.286324 4708 generic.go:334] "Generic (PLEG): container finished" podID="b26c5e08-b3b5-4f4a-9291-53a21b31f547" containerID="807e0a915b3b9f9003a5085aa9b0b1c01d8233dcbfc23e65fef49b7283eabb0e" exitCode=1 Nov 25 06:35:23 crc kubenswrapper[4708]: I1125 06:35:23.286427 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" event={"ID":"b26c5e08-b3b5-4f4a-9291-53a21b31f547","Type":"ContainerDied","Data":"807e0a915b3b9f9003a5085aa9b0b1c01d8233dcbfc23e65fef49b7283eabb0e"} Nov 25 06:35:23 crc kubenswrapper[4708]: I1125 06:35:23.286719 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" event={"ID":"b26c5e08-b3b5-4f4a-9291-53a21b31f547","Type":"ContainerStarted","Data":"88242eea158383e3bb56f76c7a8d970309b658e02901e057f4c02ebbe4404ff7"} Nov 25 06:35:23 crc kubenswrapper[4708]: I1125 06:35:23.344798 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-zqj29"] Nov 25 06:35:23 crc kubenswrapper[4708]: I1125 06:35:23.354770 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vgqw7/crc-debug-zqj29"] Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.372771 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.540356 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7fvp\" (UniqueName: \"kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp\") pod \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.540536 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host\") pod \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\" (UID: \"b26c5e08-b3b5-4f4a-9291-53a21b31f547\") " Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.540600 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host" (OuterVolumeSpecName: "host") pod "b26c5e08-b3b5-4f4a-9291-53a21b31f547" (UID: "b26c5e08-b3b5-4f4a-9291-53a21b31f547"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.541137 4708 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b26c5e08-b3b5-4f4a-9291-53a21b31f547-host\") on node \"crc\" DevicePath \"\"" Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.548721 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp" (OuterVolumeSpecName: "kube-api-access-x7fvp") pod "b26c5e08-b3b5-4f4a-9291-53a21b31f547" (UID: "b26c5e08-b3b5-4f4a-9291-53a21b31f547"). InnerVolumeSpecName "kube-api-access-x7fvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.644834 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7fvp\" (UniqueName: \"kubernetes.io/projected/b26c5e08-b3b5-4f4a-9291-53a21b31f547-kube-api-access-x7fvp\") on node \"crc\" DevicePath \"\"" Nov 25 06:35:24 crc kubenswrapper[4708]: I1125 06:35:24.902293 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b26c5e08-b3b5-4f4a-9291-53a21b31f547" path="/var/lib/kubelet/pods/b26c5e08-b3b5-4f4a-9291-53a21b31f547/volumes" Nov 25 06:35:25 crc kubenswrapper[4708]: I1125 06:35:25.302951 4708 scope.go:117] "RemoveContainer" containerID="807e0a915b3b9f9003a5085aa9b0b1c01d8233dcbfc23e65fef49b7283eabb0e" Nov 25 06:35:25 crc kubenswrapper[4708]: I1125 06:35:25.303092 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/crc-debug-zqj29" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.449460 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84bbd797c4-gvnft_236c83bd-a2bb-4d2b-b43b-212abfb9eb1f/barbican-api/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.554154 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84bbd797c4-gvnft_236c83bd-a2bb-4d2b-b43b-212abfb9eb1f/barbican-api-log/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.686949 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6db8477fbb-lpxf5_06962f63-df39-4943-8021-1e0f4d337962/barbican-keystone-listener/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.689330 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6db8477fbb-lpxf5_06962f63-df39-4943-8021-1e0f4d337962/barbican-keystone-listener-log/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.765181 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8dcdb5d95-6tkvf_92442175-4e83-4a5e-92ee-b75bfeca3f30/barbican-worker/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.908357 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8dcdb5d95-6tkvf_92442175-4e83-4a5e-92ee-b75bfeca3f30/barbican-worker-log/0.log" Nov 25 06:35:44 crc kubenswrapper[4708]: I1125 06:35:44.995807 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj_eecde679-d675-471a-9e80-a38eca4feba7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.086127 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/ceilometer-central-agent/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.125351 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/ceilometer-notification-agent/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.159593 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/sg-core/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.168812 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/proxy-httpd/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.354501 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_607ff629-5168-4560-a9bc-c4598573405b/cinder-api-log/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.445265 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_607ff629-5168-4560-a9bc-c4598573405b/cinder-api/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.515596 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fc1392ab-0788-442a-8664-1a674811e537/cinder-scheduler/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.555650 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fc1392ab-0788-442a-8664-1a674811e537/probe/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.637851 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl_87296b4d-6e9f-4206-9d0f-78dd40e51e17/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.779471 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n_9084cfbd-0f77-4a71-b265-fdbe8ac68bef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.829441 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/init/0.log" Nov 25 06:35:45 crc kubenswrapper[4708]: I1125 06:35:45.988724 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/init/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.045619 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm_555ae52e-3696-4f22-b959-d5b9407242bf/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.070551 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/dnsmasq-dns/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.251228 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0af2b359-a58e-4bc7-b497-40368df40ee1/glance-httpd/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.262769 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0af2b359-a58e-4bc7-b497-40368df40ee1/glance-log/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.380922 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6891a73d-d8e8-4a12-9136-b56912aeeb54/glance-httpd/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.419430 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6891a73d-d8e8-4a12-9136-b56912aeeb54/glance-log/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.453479 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-244cm_a0c3d4bb-310d-4de3-84d8-67f8654706b7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.614968 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4lx94_ed0b5403-c990-48b9-b3cc-8fe8455360a4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.722334 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400841-fnx8l_92abfbe3-d658-4e40-aa2b-38167fe164fd/keystone-cron/0.log" Nov 25 06:35:46 crc kubenswrapper[4708]: I1125 06:35:46.932564 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f6e8f862-6f67-426c-8e31-cb1f6baaa44e/kube-state-metrics/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.171291 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-bslnv_b24673ce-c915-4c77-aa4a-83ad370e794e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.172413 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db6d644f4-2hdvr_231fbeca-a7d9-4b28-b6ba-d1ed5ab61892/keystone-api/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.590335 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-574d769d5-pkqlp_3abea71e-bc4c-4e36-a462-bd420d9b0090/neutron-httpd/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.631675 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-574d769d5-pkqlp_3abea71e-bc4c-4e36-a462-bd420d9b0090/neutron-api/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.676834 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8/memcached/0.log" Nov 25 06:35:47 crc kubenswrapper[4708]: I1125 06:35:47.754496 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz_c88370ee-fc37-4d14-851f-87c2be9e9bf1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.119513 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12f918b9-1449-414e-abc9-723b7e31570b/nova-api-log/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.121715 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f6d2e9ec-0ff0-4db9-af17-51d9e824252d/nova-cell0-conductor-conductor/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.244836 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_6b0f9369-64c1-47a1-83ca-052238803461/nova-cell1-conductor-conductor/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.283003 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12f918b9-1449-414e-abc9-723b7e31570b/nova-api-api/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.384003 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9ad8e31-f124-4cc8-ab77-ff50df6589f1/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.468151 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-wxw56_7497674d-bdb1-427d-b4ae-cda53421f90d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.572443 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd/nova-metadata-log/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.789406 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/mysql-bootstrap/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.854464 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_815de079-46a6-4ff4-b4f3-f1e3294a21c9/nova-scheduler-scheduler/0.log" Nov 25 06:35:48 crc kubenswrapper[4708]: I1125 06:35:48.950509 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/mysql-bootstrap/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.052634 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/galera/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.072217 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/mysql-bootstrap/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.249546 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/mysql-bootstrap/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.279479 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/galera/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.305762 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd/nova-metadata-metadata/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.313124 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9c745e5e-026a-4659-8bfd-9c8939a234f4/openstackclient/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.484947 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server-init/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.501759 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bdsjf_77b982b3-1e42-4b3a-b3b9-e19f104405ed/openstack-network-exporter/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.652645 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server-init/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.682424 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovs-vswitchd/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.687761 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.714355 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-qc989_4fe9cd9f-2e87-43fa-b3c1-09560d973e2f/ovn-controller/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.942743 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_bd40d3a2-afea-42fe-93b4-72902b3046a6/openstack-network-exporter/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.979367 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-kh7hm_4b251231-3104-491a-9dc1-bac1869d716d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:49 crc kubenswrapper[4708]: I1125 06:35:49.991303 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_bd40d3a2-afea-42fe-93b4-72902b3046a6/ovn-northd/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.128051 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_131d2c08-5b83-48df-914b-1f5c28b0ba47/ovsdbserver-nb/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.134913 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_131d2c08-5b83-48df-914b-1f5c28b0ba47/openstack-network-exporter/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.172583 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4c903b56-d8a2-4522-97d1-4b2333be4929/openstack-network-exporter/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.225125 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4c903b56-d8a2-4522-97d1-4b2333be4929/ovsdbserver-sb/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.347026 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8ff6575b-8w7c4_17af25e9-9b9e-4175-b35c-18fe0b9e0c36/placement-api/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.406001 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/setup-container/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.436919 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8ff6575b-8w7c4_17af25e9-9b9e-4175-b35c-18fe0b9e0c36/placement-log/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.597323 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/rabbitmq/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.615275 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/setup-container/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.620607 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/setup-container/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.802368 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/setup-container/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.802586 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/rabbitmq/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.847793 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6_1ffad4ac-de8a-4f8c-8409-20b8553a4259/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:50 crc kubenswrapper[4708]: I1125 06:35:50.983233 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-d6wtp_19981de6-1c24-415c-9d45-c2d31d8b54fd/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.031455 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-72fph_50e3a68c-2f8a-48c7-a330-aa646ee174dd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.083792 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6k7zh_1ef0663b-adb0-4192-a17a-1643899f466f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.157595 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-t7t6t_a2cb620c-a539-4928-9a66-6c08cf197798/ssh-known-hosts-edpm-deployment/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.314898 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b4dc74dc-ktwwc_34548024-cb82-4fa9-ace7-9e41b3ab2c1d/proxy-server/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.363235 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b4dc74dc-ktwwc_34548024-cb82-4fa9-ace7-9e41b3ab2c1d/proxy-httpd/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.380988 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-p5f7m_14772fed-d981-4b42-ac9b-24dd6ce02b06/swift-ring-rebalance/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.516651 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-auditor/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.530680 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-reaper/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.553021 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-replicator/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.588820 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-server/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.677913 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-auditor/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.705791 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-server/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.715465 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-updater/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.726408 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-replicator/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.772841 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-auditor/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.848649 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-expirer/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.884964 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-server/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.898882 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-updater/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.900264 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-replicator/0.log" Nov 25 06:35:51 crc kubenswrapper[4708]: I1125 06:35:51.966217 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/rsync/0.log" Nov 25 06:35:52 crc kubenswrapper[4708]: I1125 06:35:52.087175 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/swift-recon-cron/0.log" Nov 25 06:35:52 crc kubenswrapper[4708]: I1125 06:35:52.132020 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q_180c2306-33c0-4aed-93f8-3c99603f678b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:35:52 crc kubenswrapper[4708]: I1125 06:35:52.195423 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_695e5c80-15cf-4dc5-b286-cb10e12e9fee/tempest-tests-tempest-tests-runner/0.log" Nov 25 06:35:52 crc kubenswrapper[4708]: I1125 06:35:52.262016 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_655d4e8f-367a-475f-8363-1506473bb0d4/test-operator-logs-container/0.log" Nov 25 06:35:52 crc kubenswrapper[4708]: I1125 06:35:52.309180 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4w55g_ec676a79-152d-4660-854e-5e447bd52ff2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:36:11 crc kubenswrapper[4708]: I1125 06:36:11.858032 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-kvrrg_8cb5eab2-d137-4aeb-9ca5-59139c67a2de/kube-rbac-proxy/0.log" Nov 25 06:36:11 crc kubenswrapper[4708]: I1125 06:36:11.919874 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-kvrrg_8cb5eab2-d137-4aeb-9ca5-59139c67a2de/manager/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.051500 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.235897 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.243503 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.262287 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.410134 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.424461 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.456793 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/extract/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.605847 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-b7hd2_13a9d474-f528-4913-ada4-313db91a37d5/kube-rbac-proxy/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.639121 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-b7hd2_13a9d474-f528-4913-ada4-313db91a37d5/manager/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.743497 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-j2vk8_fd2a735a-3219-45d8-a689-1a8722923d8d/kube-rbac-proxy/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.792036 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-j2vk8_fd2a735a-3219-45d8-a689-1a8722923d8d/manager/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.820385 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-df8zr_eef4d5f6-a70a-4b08-8836-3dde6ab0da36/kube-rbac-proxy/0.log" Nov 25 06:36:12 crc kubenswrapper[4708]: I1125 06:36:12.995074 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-df8zr_eef4d5f6-a70a-4b08-8836-3dde6ab0da36/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.017594 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-9knwb_e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.027249 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-9knwb_e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.194292 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j98rt_b752b0d3-bffb-4322-9e11-d2a9451c4f39/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.196794 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j98rt_b752b0d3-bffb-4322-9e11-d2a9451c4f39/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.338194 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-j5g96_49cb03d2-7430-4f92-b82b-7e4b7ad9faee/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.398223 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-8xsh6_1fd35d81-f228-4c45-895a-95feb523ef1f/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.539548 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-j5g96_49cb03d2-7430-4f92-b82b-7e4b7ad9faee/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.550683 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-8xsh6_1fd35d81-f228-4c45-895a-95feb523ef1f/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.601012 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-9wnpj_1e34dcae-a471-4c9c-83f2-43d551516520/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.769512 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-9wnpj_1e34dcae-a471-4c9c-83f2-43d551516520/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.783119 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-6zc24_b2240efc-4688-4d79-a19d-ea3801742494/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.799963 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-6zc24_b2240efc-4688-4d79-a19d-ea3801742494/kube-rbac-proxy/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.958806 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-x4xzv_856fc8ab-76a7-4ed4-ab1e-31dad3985982/manager/0.log" Nov 25 06:36:13 crc kubenswrapper[4708]: I1125 06:36:13.990126 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-x4xzv_856fc8ab-76a7-4ed4-ab1e-31dad3985982/kube-rbac-proxy/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.121253 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-c8c5l_17ef75be-981e-4579-ae1b-aa7e4c4b8918/kube-rbac-proxy/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.189933 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-c8c5l_17ef75be-981e-4579-ae1b-aa7e4c4b8918/manager/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.229851 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-qjcbh_dcbf8314-b18c-43ce-8a2e-400b9da62852/kube-rbac-proxy/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.356671 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.356720 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.409277 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-qjcbh_dcbf8314-b18c-43ce-8a2e-400b9da62852/manager/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.467103 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rmmvr_6af5f14d-1c9e-4985-ba46-3f4b193ac96d/kube-rbac-proxy/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.468138 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rmmvr_6af5f14d-1c9e-4985-ba46-3f4b193ac96d/manager/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.633408 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gk8ft_3d11487d-d74d-428d-b18b-800238416691/kube-rbac-proxy/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.664864 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gk8ft_3d11487d-d74d-428d-b18b-800238416691/manager/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.931723 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b567956b5-lb9hh_5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1/operator/0.log" Nov 25 06:36:14 crc kubenswrapper[4708]: I1125 06:36:14.978737 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zkbgb_7dc7a5c4-2123-4ca4-ad42-364233f79bb5/registry-server/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.300617 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-gnn9s_a7c6b74d-a672-4349-b9f7-52a726bc8ec9/kube-rbac-proxy/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.384310 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-gnn9s_a7c6b74d-a672-4349-b9f7-52a726bc8ec9/manager/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.484809 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-vzz64_1064823e-17fb-42bc-b68e-cfab481fbf8c/kube-rbac-proxy/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.498402 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-vzz64_1064823e-17fb-42bc-b68e-cfab481fbf8c/manager/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.567927 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cd5954d9-c7dbb_af33b085-a912-42a9-88ba-6b7e8d27ec8d/manager/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.590937 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mc72h_54a0ed25-ded8-471d-acdd-b34f75347701/operator/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.742021 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-d64tb_83077424-8cf5-48ae-b8b2-d0b94fb650e6/manager/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.754045 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-d64tb_83077424-8cf5-48ae-b8b2-d0b94fb650e6/kube-rbac-proxy/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.777787 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-phd66_7d60669a-b373-49bf-b65f-88b3e0d48d87/kube-rbac-proxy/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.936238 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-phd66_7d60669a-b373-49bf-b65f-88b3e0d48d87/manager/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.970093 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-z9kvf_ed82966d-e1bb-448f-be61-091fcb5f3f2b/kube-rbac-proxy/0.log" Nov 25 06:36:15 crc kubenswrapper[4708]: I1125 06:36:15.999821 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-z9kvf_ed82966d-e1bb-448f-be61-091fcb5f3f2b/manager/0.log" Nov 25 06:36:16 crc kubenswrapper[4708]: I1125 06:36:16.106708 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-hhgb4_cb75b241-90c4-45ee-9473-864b6b9f8705/manager/0.log" Nov 25 06:36:16 crc kubenswrapper[4708]: I1125 06:36:16.109540 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-hhgb4_cb75b241-90c4-45ee-9473-864b6b9f8705/kube-rbac-proxy/0.log" Nov 25 06:36:31 crc kubenswrapper[4708]: I1125 06:36:31.880731 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-642cw_e1dcf9d6-33d7-4482-b24f-990de83f2fa1/control-plane-machine-set-operator/0.log" Nov 25 06:36:32 crc kubenswrapper[4708]: I1125 06:36:32.034361 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f98lb_e41814f6-cdb7-4460-bc75-618b5fff39df/kube-rbac-proxy/0.log" Nov 25 06:36:32 crc kubenswrapper[4708]: I1125 06:36:32.051959 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f98lb_e41814f6-cdb7-4460-bc75-618b5fff39df/machine-api-operator/0.log" Nov 25 06:36:43 crc kubenswrapper[4708]: I1125 06:36:43.572111 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pfp8b_8775bc9d-3d1d-471a-b6af-c057f25a7f12/cert-manager-controller/0.log" Nov 25 06:36:43 crc kubenswrapper[4708]: I1125 06:36:43.696807 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-t4xc2_6cd001d7-f756-4a83-8d4b-0bc9c5198b0a/cert-manager-cainjector/0.log" Nov 25 06:36:43 crc kubenswrapper[4708]: I1125 06:36:43.745391 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7dc5l_d043f20a-ee01-40c7-86a5-ff37f52d7f07/cert-manager-webhook/0.log" Nov 25 06:36:44 crc kubenswrapper[4708]: I1125 06:36:44.357037 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:36:44 crc kubenswrapper[4708]: I1125 06:36:44.357117 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.525254 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-w982m_d6725355-22e6-4236-94e9-ce851f0d8e29/nmstate-console-plugin/0.log" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.699820 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gghd9_7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7/nmstate-handler/0.log" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.730798 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kjddh_f7336ae4-7e4b-48a3-8fbd-618e0491a2ff/kube-rbac-proxy/0.log" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.730963 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kjddh_f7336ae4-7e4b-48a3-8fbd-618e0491a2ff/nmstate-metrics/0.log" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.871860 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-dvth5_daa36bec-fbbf-44f3-8785-8e09474d7add/nmstate-operator/0.log" Nov 25 06:36:54 crc kubenswrapper[4708]: I1125 06:36:54.888856 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-2xlzf_03a47fb5-3577-45b2-83b2-cedcd799eea0/nmstate-webhook/0.log" Nov 25 06:37:06 crc kubenswrapper[4708]: I1125 06:37:06.939791 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bkbdq_56eb3e20-e964-4e88-a2f7-8f07c3e8d40f/kube-rbac-proxy/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.084713 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bkbdq_56eb3e20-e964-4e88-a2f7-8f07c3e8d40f/controller/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.127569 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.298554 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.319633 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.341242 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.357164 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.516490 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.523923 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.530002 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.542205 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.749281 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.753702 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.758046 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.790815 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/controller/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.917807 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/frr-metrics/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.942809 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/kube-rbac-proxy/0.log" Nov 25 06:37:07 crc kubenswrapper[4708]: I1125 06:37:07.961452 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/kube-rbac-proxy-frr/0.log" Nov 25 06:37:08 crc kubenswrapper[4708]: I1125 06:37:08.187494 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-5fbpc_f3e32da4-a772-4f5b-b91a-fbd4f5405a94/frr-k8s-webhook-server/0.log" Nov 25 06:37:08 crc kubenswrapper[4708]: I1125 06:37:08.189751 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/reloader/0.log" Nov 25 06:37:08 crc kubenswrapper[4708]: I1125 06:37:08.363475 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5885d8f89f-gjghz_93f12b13-309e-4a76-9f10-c1ca9a0545eb/manager/0.log" Nov 25 06:37:08 crc kubenswrapper[4708]: I1125 06:37:08.563835 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66489dff8c-hbstx_9ae15378-5b01-442c-a1b3-52d25e474875/webhook-server/0.log" Nov 25 06:37:08 crc kubenswrapper[4708]: I1125 06:37:08.608972 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dg7zk_fb80d906-e252-4e84-95a3-8c63ef010175/kube-rbac-proxy/0.log" Nov 25 06:37:09 crc kubenswrapper[4708]: I1125 06:37:09.173445 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/frr/0.log" Nov 25 06:37:09 crc kubenswrapper[4708]: I1125 06:37:09.179139 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dg7zk_fb80d906-e252-4e84-95a3-8c63ef010175/speaker/0.log" Nov 25 06:37:14 crc kubenswrapper[4708]: I1125 06:37:14.356871 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:37:14 crc kubenswrapper[4708]: I1125 06:37:14.357764 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:37:14 crc kubenswrapper[4708]: I1125 06:37:14.357844 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:37:14 crc kubenswrapper[4708]: I1125 06:37:14.358984 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:37:14 crc kubenswrapper[4708]: I1125 06:37:14.359064 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30" gracePeriod=600 Nov 25 06:37:15 crc kubenswrapper[4708]: I1125 06:37:15.254750 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30" exitCode=0 Nov 25 06:37:15 crc kubenswrapper[4708]: I1125 06:37:15.254815 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30"} Nov 25 06:37:15 crc kubenswrapper[4708]: I1125 06:37:15.255451 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4"} Nov 25 06:37:15 crc kubenswrapper[4708]: I1125 06:37:15.255482 4708 scope.go:117] "RemoveContainer" containerID="31a44ad36120270362a0ae671e6767581e4fd4519827c0e291e5998e629c11c4" Nov 25 06:37:19 crc kubenswrapper[4708]: I1125 06:37:19.850629 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:37:19 crc kubenswrapper[4708]: I1125 06:37:19.936097 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:37:19 crc kubenswrapper[4708]: I1125 06:37:19.962323 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.013357 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.142927 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/extract/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.143297 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.159343 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.308204 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.458346 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.465616 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.467991 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.624754 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.629681 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:37:20 crc kubenswrapper[4708]: I1125 06:37:20.801179 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.041443 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.055806 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.120278 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.155446 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/registry-server/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.228483 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.237416 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.429946 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.624708 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.632145 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.652728 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.732023 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/registry-server/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.799362 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.799783 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.856787 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/extract/0.log" Nov 25 06:37:21 crc kubenswrapper[4708]: I1125 06:37:21.955437 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qlzlc_7fa8126c-8308-4ce3-8529-70b28a606428/marketplace-operator/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.037451 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.194758 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.194863 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.196646 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.331936 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.364100 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.468277 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/registry-server/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.531270 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.673549 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.680352 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.695596 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.805498 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:37:22 crc kubenswrapper[4708]: I1125 06:37:22.831767 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:37:23 crc kubenswrapper[4708]: I1125 06:37:23.205495 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/registry-server/0.log" Nov 25 06:38:46 crc kubenswrapper[4708]: I1125 06:38:46.111510 4708 generic.go:334] "Generic (PLEG): container finished" podID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerID="5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766" exitCode=0 Nov 25 06:38:46 crc kubenswrapper[4708]: I1125 06:38:46.111611 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" event={"ID":"574fa4f1-02e8-4576-ac43-7a97950d5431","Type":"ContainerDied","Data":"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766"} Nov 25 06:38:46 crc kubenswrapper[4708]: I1125 06:38:46.112970 4708 scope.go:117] "RemoveContainer" containerID="5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766" Nov 25 06:38:46 crc kubenswrapper[4708]: I1125 06:38:46.181758 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vgqw7_must-gather-hgfkx_574fa4f1-02e8-4576-ac43-7a97950d5431/gather/0.log" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.176898 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vgqw7/must-gather-hgfkx"] Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.177794 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="copy" containerID="cri-o://43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821" gracePeriod=2 Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.183798 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vgqw7/must-gather-hgfkx"] Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.582193 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vgqw7_must-gather-hgfkx_574fa4f1-02e8-4576-ac43-7a97950d5431/copy/0.log" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.583155 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.665184 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p4gn\" (UniqueName: \"kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn\") pod \"574fa4f1-02e8-4576-ac43-7a97950d5431\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.665447 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output\") pod \"574fa4f1-02e8-4576-ac43-7a97950d5431\" (UID: \"574fa4f1-02e8-4576-ac43-7a97950d5431\") " Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.672293 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn" (OuterVolumeSpecName: "kube-api-access-6p4gn") pod "574fa4f1-02e8-4576-ac43-7a97950d5431" (UID: "574fa4f1-02e8-4576-ac43-7a97950d5431"). InnerVolumeSpecName "kube-api-access-6p4gn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.768217 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p4gn\" (UniqueName: \"kubernetes.io/projected/574fa4f1-02e8-4576-ac43-7a97950d5431-kube-api-access-6p4gn\") on node \"crc\" DevicePath \"\"" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.782632 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "574fa4f1-02e8-4576-ac43-7a97950d5431" (UID: "574fa4f1-02e8-4576-ac43-7a97950d5431"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.869423 4708 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/574fa4f1-02e8-4576-ac43-7a97950d5431-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 06:38:54 crc kubenswrapper[4708]: I1125 06:38:54.907105 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" path="/var/lib/kubelet/pods/574fa4f1-02e8-4576-ac43-7a97950d5431/volumes" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.195672 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vgqw7_must-gather-hgfkx_574fa4f1-02e8-4576-ac43-7a97950d5431/copy/0.log" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.196125 4708 generic.go:334] "Generic (PLEG): container finished" podID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerID="43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821" exitCode=143 Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.196209 4708 scope.go:117] "RemoveContainer" containerID="43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.196241 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vgqw7/must-gather-hgfkx" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.224582 4708 scope.go:117] "RemoveContainer" containerID="5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.276569 4708 scope.go:117] "RemoveContainer" containerID="43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821" Nov 25 06:38:55 crc kubenswrapper[4708]: E1125 06:38:55.277413 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821\": container with ID starting with 43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821 not found: ID does not exist" containerID="43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.277464 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821"} err="failed to get container status \"43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821\": rpc error: code = NotFound desc = could not find container \"43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821\": container with ID starting with 43dc0cb9a5b7da693d05608fff7102acf3ad371a1d9432872e1383d813f6f821 not found: ID does not exist" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.277496 4708 scope.go:117] "RemoveContainer" containerID="5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766" Nov 25 06:38:55 crc kubenswrapper[4708]: E1125 06:38:55.278126 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766\": container with ID starting with 5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766 not found: ID does not exist" containerID="5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766" Nov 25 06:38:55 crc kubenswrapper[4708]: I1125 06:38:55.278188 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766"} err="failed to get container status \"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766\": rpc error: code = NotFound desc = could not find container \"5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766\": container with ID starting with 5f1bb315c12468ec49e083ae67acf15817feae455ec302ba30a35d07260c0766 not found: ID does not exist" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.855668 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:38:57 crc kubenswrapper[4708]: E1125 06:38:57.856650 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b26c5e08-b3b5-4f4a-9291-53a21b31f547" containerName="container-00" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856665 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b26c5e08-b3b5-4f4a-9291-53a21b31f547" containerName="container-00" Nov 25 06:38:57 crc kubenswrapper[4708]: E1125 06:38:57.856676 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="copy" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856681 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="copy" Nov 25 06:38:57 crc kubenswrapper[4708]: E1125 06:38:57.856694 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="gather" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856701 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="gather" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856927 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="copy" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856939 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="574fa4f1-02e8-4576-ac43-7a97950d5431" containerName="gather" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.856946 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b26c5e08-b3b5-4f4a-9291-53a21b31f547" containerName="container-00" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.858549 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.875854 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.942339 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.942472 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:57 crc kubenswrapper[4708]: I1125 06:38:57.942496 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crs4d\" (UniqueName: \"kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.043929 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.044045 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crs4d\" (UniqueName: \"kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.044402 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.044548 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.044930 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.099093 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crs4d\" (UniqueName: \"kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d\") pod \"redhat-operators-jvwx6\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.178268 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:38:58 crc kubenswrapper[4708]: I1125 06:38:58.598593 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:38:59 crc kubenswrapper[4708]: I1125 06:38:59.239809 4708 generic.go:334] "Generic (PLEG): container finished" podID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerID="b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c" exitCode=0 Nov 25 06:38:59 crc kubenswrapper[4708]: I1125 06:38:59.239930 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerDied","Data":"b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c"} Nov 25 06:38:59 crc kubenswrapper[4708]: I1125 06:38:59.240200 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerStarted","Data":"490659c02d78ddf051d77debd4ba139ff5914b57a2f4c6c2fca08ee98221ad95"} Nov 25 06:39:00 crc kubenswrapper[4708]: I1125 06:39:00.253753 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerStarted","Data":"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904"} Nov 25 06:39:01 crc kubenswrapper[4708]: I1125 06:39:01.271165 4708 generic.go:334] "Generic (PLEG): container finished" podID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerID="3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904" exitCode=0 Nov 25 06:39:01 crc kubenswrapper[4708]: I1125 06:39:01.271283 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerDied","Data":"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904"} Nov 25 06:39:02 crc kubenswrapper[4708]: I1125 06:39:02.282768 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerStarted","Data":"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce"} Nov 25 06:39:02 crc kubenswrapper[4708]: I1125 06:39:02.307253 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jvwx6" podStartSLOduration=2.786588153 podStartE2EDuration="5.307227882s" podCreationTimestamp="2025-11-25 06:38:57 +0000 UTC" firstStartedPulling="2025-11-25 06:38:59.241585914 +0000 UTC m=+3480.650419300" lastFinishedPulling="2025-11-25 06:39:01.762225644 +0000 UTC m=+3483.171059029" observedRunningTime="2025-11-25 06:39:02.300199092 +0000 UTC m=+3483.709032478" watchObservedRunningTime="2025-11-25 06:39:02.307227882 +0000 UTC m=+3483.716061268" Nov 25 06:39:08 crc kubenswrapper[4708]: I1125 06:39:08.178913 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:08 crc kubenswrapper[4708]: I1125 06:39:08.179587 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:08 crc kubenswrapper[4708]: I1125 06:39:08.222967 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:08 crc kubenswrapper[4708]: I1125 06:39:08.371063 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:08 crc kubenswrapper[4708]: I1125 06:39:08.457761 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:39:10 crc kubenswrapper[4708]: I1125 06:39:10.355296 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jvwx6" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="registry-server" containerID="cri-o://6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce" gracePeriod=2 Nov 25 06:39:11 crc kubenswrapper[4708]: I1125 06:39:11.877956 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.061798 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crs4d\" (UniqueName: \"kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d\") pod \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.062062 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities\") pod \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.062250 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content\") pod \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\" (UID: \"e10f1c79-3616-4583-9c7b-cb5c7b701dee\") " Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.063127 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities" (OuterVolumeSpecName: "utilities") pod "e10f1c79-3616-4583-9c7b-cb5c7b701dee" (UID: "e10f1c79-3616-4583-9c7b-cb5c7b701dee"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.069219 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d" (OuterVolumeSpecName: "kube-api-access-crs4d") pod "e10f1c79-3616-4583-9c7b-cb5c7b701dee" (UID: "e10f1c79-3616-4583-9c7b-cb5c7b701dee"). InnerVolumeSpecName "kube-api-access-crs4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.144759 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e10f1c79-3616-4583-9c7b-cb5c7b701dee" (UID: "e10f1c79-3616-4583-9c7b-cb5c7b701dee"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.164884 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.164934 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e10f1c79-3616-4583-9c7b-cb5c7b701dee-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.164950 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crs4d\" (UniqueName: \"kubernetes.io/projected/e10f1c79-3616-4583-9c7b-cb5c7b701dee-kube-api-access-crs4d\") on node \"crc\" DevicePath \"\"" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.378106 4708 generic.go:334] "Generic (PLEG): container finished" podID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerID="6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce" exitCode=0 Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.378175 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerDied","Data":"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce"} Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.378235 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jvwx6" event={"ID":"e10f1c79-3616-4583-9c7b-cb5c7b701dee","Type":"ContainerDied","Data":"490659c02d78ddf051d77debd4ba139ff5914b57a2f4c6c2fca08ee98221ad95"} Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.378262 4708 scope.go:117] "RemoveContainer" containerID="6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.378449 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jvwx6" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.397660 4708 scope.go:117] "RemoveContainer" containerID="3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.412030 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.418413 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jvwx6"] Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.426256 4708 scope.go:117] "RemoveContainer" containerID="b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.451179 4708 scope.go:117] "RemoveContainer" containerID="6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce" Nov 25 06:39:12 crc kubenswrapper[4708]: E1125 06:39:12.451666 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce\": container with ID starting with 6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce not found: ID does not exist" containerID="6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.451705 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce"} err="failed to get container status \"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce\": rpc error: code = NotFound desc = could not find container \"6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce\": container with ID starting with 6fded167908233663ff51683a2e1a2efe91804875648567d23afbbae689777ce not found: ID does not exist" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.451730 4708 scope.go:117] "RemoveContainer" containerID="3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904" Nov 25 06:39:12 crc kubenswrapper[4708]: E1125 06:39:12.452209 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904\": container with ID starting with 3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904 not found: ID does not exist" containerID="3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.452252 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904"} err="failed to get container status \"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904\": rpc error: code = NotFound desc = could not find container \"3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904\": container with ID starting with 3f3e57dc13f7a0ccc453c18d3e83517792bc624ed7fa0cc7985f1ce70eb2f904 not found: ID does not exist" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.452274 4708 scope.go:117] "RemoveContainer" containerID="b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c" Nov 25 06:39:12 crc kubenswrapper[4708]: E1125 06:39:12.452677 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c\": container with ID starting with b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c not found: ID does not exist" containerID="b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.452730 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c"} err="failed to get container status \"b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c\": rpc error: code = NotFound desc = could not find container \"b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c\": container with ID starting with b5fd70ffe9ee0d0d73ee7739579028ff61f983c85e884f93abd549e9bb150c5c not found: ID does not exist" Nov 25 06:39:12 crc kubenswrapper[4708]: I1125 06:39:12.904370 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" path="/var/lib/kubelet/pods/e10f1c79-3616-4583-9c7b-cb5c7b701dee/volumes" Nov 25 06:39:14 crc kubenswrapper[4708]: I1125 06:39:14.357381 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:39:14 crc kubenswrapper[4708]: I1125 06:39:14.357463 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:39:44 crc kubenswrapper[4708]: I1125 06:39:44.356795 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:39:44 crc kubenswrapper[4708]: I1125 06:39:44.357433 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.357335 4708 patch_prober.go:28] interesting pod/machine-config-daemon-lh62q container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.358122 4708 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.358192 4708 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.358875 4708 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4"} pod="openshift-machine-config-operator/machine-config-daemon-lh62q" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.358947 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" containerName="machine-config-daemon" containerID="cri-o://12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" gracePeriod=600 Nov 25 06:40:14 crc kubenswrapper[4708]: E1125 06:40:14.477670 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.969149 4708 generic.go:334] "Generic (PLEG): container finished" podID="f24da5a0-0c94-4513-9c44-84818f240046" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" exitCode=0 Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.969202 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerDied","Data":"12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4"} Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.969257 4708 scope.go:117] "RemoveContainer" containerID="32eeb928f62d737c87f5254dd94f6a6836c54d0176ee48dba8dc5d1ae1ea8e30" Nov 25 06:40:14 crc kubenswrapper[4708]: I1125 06:40:14.970029 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:40:14 crc kubenswrapper[4708]: E1125 06:40:14.970339 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:40:27 crc kubenswrapper[4708]: I1125 06:40:27.894160 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:40:27 crc kubenswrapper[4708]: E1125 06:40:27.895095 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:40:39 crc kubenswrapper[4708]: I1125 06:40:39.893854 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:40:39 crc kubenswrapper[4708]: E1125 06:40:39.894943 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:40:51 crc kubenswrapper[4708]: I1125 06:40:51.894011 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:40:51 crc kubenswrapper[4708]: E1125 06:40:51.895071 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.100632 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4cx2b/must-gather-tsjj4"] Nov 25 06:40:56 crc kubenswrapper[4708]: E1125 06:40:56.102146 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="extract-utilities" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.102165 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="extract-utilities" Nov 25 06:40:56 crc kubenswrapper[4708]: E1125 06:40:56.102192 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="registry-server" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.102198 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="registry-server" Nov 25 06:40:56 crc kubenswrapper[4708]: E1125 06:40:56.102233 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="extract-content" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.102239 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="extract-content" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.102454 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="e10f1c79-3616-4583-9c7b-cb5c7b701dee" containerName="registry-server" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.103991 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.106171 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4cx2b"/"openshift-service-ca.crt" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.106325 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmvvc\" (UniqueName: \"kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.106430 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4cx2b"/"default-dockercfg-bkg29" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.106759 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.106820 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4cx2b"/"kube-root-ca.crt" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.124065 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4cx2b/must-gather-tsjj4"] Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.208868 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmvvc\" (UniqueName: \"kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.208966 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.209443 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.228210 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmvvc\" (UniqueName: \"kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc\") pod \"must-gather-tsjj4\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.426508 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:40:56 crc kubenswrapper[4708]: I1125 06:40:56.878057 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4cx2b/must-gather-tsjj4"] Nov 25 06:40:57 crc kubenswrapper[4708]: I1125 06:40:57.405811 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" event={"ID":"e1f20800-2884-4fb3-ade5-bc1860682bea","Type":"ContainerStarted","Data":"3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec"} Nov 25 06:40:57 crc kubenswrapper[4708]: I1125 06:40:57.407384 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" event={"ID":"e1f20800-2884-4fb3-ade5-bc1860682bea","Type":"ContainerStarted","Data":"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b"} Nov 25 06:40:57 crc kubenswrapper[4708]: I1125 06:40:57.407459 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" event={"ID":"e1f20800-2884-4fb3-ade5-bc1860682bea","Type":"ContainerStarted","Data":"10d2adc365d504a7fc2a1b43a5cbb09b1b7709d0f2ebc629ee36b4cc5f878126"} Nov 25 06:40:57 crc kubenswrapper[4708]: I1125 06:40:57.426638 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" podStartSLOduration=1.426624885 podStartE2EDuration="1.426624885s" podCreationTimestamp="2025-11-25 06:40:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 06:40:57.42036682 +0000 UTC m=+3598.829200206" watchObservedRunningTime="2025-11-25 06:40:57.426624885 +0000 UTC m=+3598.835458272" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.430234 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-9v5pv"] Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.431699 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.503398 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl7tr\" (UniqueName: \"kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.503591 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.605637 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl7tr\" (UniqueName: \"kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.605866 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.606024 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.622756 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl7tr\" (UniqueName: \"kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr\") pod \"crc-debug-9v5pv\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:00 crc kubenswrapper[4708]: I1125 06:41:00.747086 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:01 crc kubenswrapper[4708]: I1125 06:41:01.450304 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" event={"ID":"70878a9a-6223-43b0-9a14-5fc7d9c4fb85","Type":"ContainerStarted","Data":"7098dc40bd9845ff9772139c42740cb807d895964f4b39e1c3a30113a8ea19a8"} Nov 25 06:41:01 crc kubenswrapper[4708]: I1125 06:41:01.451023 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" event={"ID":"70878a9a-6223-43b0-9a14-5fc7d9c4fb85","Type":"ContainerStarted","Data":"20657e6b755c0be1a0374da312a75eb505d9d50a85e227627058e034d1305abf"} Nov 25 06:41:01 crc kubenswrapper[4708]: I1125 06:41:01.471400 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" podStartSLOduration=1.471375406 podStartE2EDuration="1.471375406s" podCreationTimestamp="2025-11-25 06:41:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 06:41:01.471041516 +0000 UTC m=+3602.879874902" watchObservedRunningTime="2025-11-25 06:41:01.471375406 +0000 UTC m=+3602.880208792" Nov 25 06:41:02 crc kubenswrapper[4708]: I1125 06:41:02.538890 4708 scope.go:117] "RemoveContainer" containerID="731b10a3e32d4f5f25d2c4a241a9de57dd94494e2427f32f2e7756e3aa445dc6" Nov 25 06:41:03 crc kubenswrapper[4708]: I1125 06:41:03.893334 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:41:03 crc kubenswrapper[4708]: E1125 06:41:03.893861 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:41:10 crc kubenswrapper[4708]: I1125 06:41:10.530484 4708 generic.go:334] "Generic (PLEG): container finished" podID="70878a9a-6223-43b0-9a14-5fc7d9c4fb85" containerID="7098dc40bd9845ff9772139c42740cb807d895964f4b39e1c3a30113a8ea19a8" exitCode=0 Nov 25 06:41:10 crc kubenswrapper[4708]: I1125 06:41:10.530579 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" event={"ID":"70878a9a-6223-43b0-9a14-5fc7d9c4fb85","Type":"ContainerDied","Data":"7098dc40bd9845ff9772139c42740cb807d895964f4b39e1c3a30113a8ea19a8"} Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.631251 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.647067 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl7tr\" (UniqueName: \"kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr\") pod \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.653457 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr" (OuterVolumeSpecName: "kube-api-access-kl7tr") pod "70878a9a-6223-43b0-9a14-5fc7d9c4fb85" (UID: "70878a9a-6223-43b0-9a14-5fc7d9c4fb85"). InnerVolumeSpecName "kube-api-access-kl7tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.668885 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-9v5pv"] Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.674796 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-9v5pv"] Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.748617 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host\") pod \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\" (UID: \"70878a9a-6223-43b0-9a14-5fc7d9c4fb85\") " Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.748728 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host" (OuterVolumeSpecName: "host") pod "70878a9a-6223-43b0-9a14-5fc7d9c4fb85" (UID: "70878a9a-6223-43b0-9a14-5fc7d9c4fb85"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.749148 4708 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-host\") on node \"crc\" DevicePath \"\"" Nov 25 06:41:11 crc kubenswrapper[4708]: I1125 06:41:11.749172 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl7tr\" (UniqueName: \"kubernetes.io/projected/70878a9a-6223-43b0-9a14-5fc7d9c4fb85-kube-api-access-kl7tr\") on node \"crc\" DevicePath \"\"" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.548788 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20657e6b755c0be1a0374da312a75eb505d9d50a85e227627058e034d1305abf" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.548840 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-9v5pv" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.820548 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-vmr8x"] Nov 25 06:41:12 crc kubenswrapper[4708]: E1125 06:41:12.820934 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70878a9a-6223-43b0-9a14-5fc7d9c4fb85" containerName="container-00" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.820948 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="70878a9a-6223-43b0-9a14-5fc7d9c4fb85" containerName="container-00" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.821149 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="70878a9a-6223-43b0-9a14-5fc7d9c4fb85" containerName="container-00" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.821777 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.871835 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.872162 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbz2\" (UniqueName: \"kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.901921 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70878a9a-6223-43b0-9a14-5fc7d9c4fb85" path="/var/lib/kubelet/pods/70878a9a-6223-43b0-9a14-5fc7d9c4fb85/volumes" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.975719 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbz2\" (UniqueName: \"kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.975902 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.976011 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:12 crc kubenswrapper[4708]: I1125 06:41:12.991635 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbz2\" (UniqueName: \"kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2\") pod \"crc-debug-vmr8x\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.138873 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.560217 4708 generic.go:334] "Generic (PLEG): container finished" podID="24b001fe-8c17-4769-acd2-cbdb94abf51d" containerID="7d2e7e66177da602106112a4e139d968f074c74db95e0365aa954286a046d5f3" exitCode=1 Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.560268 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" event={"ID":"24b001fe-8c17-4769-acd2-cbdb94abf51d","Type":"ContainerDied","Data":"7d2e7e66177da602106112a4e139d968f074c74db95e0365aa954286a046d5f3"} Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.560349 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" event={"ID":"24b001fe-8c17-4769-acd2-cbdb94abf51d","Type":"ContainerStarted","Data":"8ca51b97add3919a8e2bed42b589412a5a6b98066d85bd9f4765bda3e25e845a"} Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.601826 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-vmr8x"] Nov 25 06:41:13 crc kubenswrapper[4708]: I1125 06:41:13.608044 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4cx2b/crc-debug-vmr8x"] Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.645432 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.713895 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host\") pod \"24b001fe-8c17-4769-acd2-cbdb94abf51d\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.714019 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host" (OuterVolumeSpecName: "host") pod "24b001fe-8c17-4769-acd2-cbdb94abf51d" (UID: "24b001fe-8c17-4769-acd2-cbdb94abf51d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.714258 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjbz2\" (UniqueName: \"kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2\") pod \"24b001fe-8c17-4769-acd2-cbdb94abf51d\" (UID: \"24b001fe-8c17-4769-acd2-cbdb94abf51d\") " Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.715175 4708 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/24b001fe-8c17-4769-acd2-cbdb94abf51d-host\") on node \"crc\" DevicePath \"\"" Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.722360 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2" (OuterVolumeSpecName: "kube-api-access-fjbz2") pod "24b001fe-8c17-4769-acd2-cbdb94abf51d" (UID: "24b001fe-8c17-4769-acd2-cbdb94abf51d"). InnerVolumeSpecName "kube-api-access-fjbz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.817264 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjbz2\" (UniqueName: \"kubernetes.io/projected/24b001fe-8c17-4769-acd2-cbdb94abf51d-kube-api-access-fjbz2\") on node \"crc\" DevicePath \"\"" Nov 25 06:41:14 crc kubenswrapper[4708]: I1125 06:41:14.904276 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24b001fe-8c17-4769-acd2-cbdb94abf51d" path="/var/lib/kubelet/pods/24b001fe-8c17-4769-acd2-cbdb94abf51d/volumes" Nov 25 06:41:15 crc kubenswrapper[4708]: I1125 06:41:15.575758 4708 scope.go:117] "RemoveContainer" containerID="7d2e7e66177da602106112a4e139d968f074c74db95e0365aa954286a046d5f3" Nov 25 06:41:15 crc kubenswrapper[4708]: I1125 06:41:15.575787 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/crc-debug-vmr8x" Nov 25 06:41:15 crc kubenswrapper[4708]: I1125 06:41:15.893625 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:41:15 crc kubenswrapper[4708]: E1125 06:41:15.894088 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:41:26 crc kubenswrapper[4708]: I1125 06:41:26.893983 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:41:26 crc kubenswrapper[4708]: E1125 06:41:26.894926 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:41:40 crc kubenswrapper[4708]: I1125 06:41:40.895998 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:41:40 crc kubenswrapper[4708]: E1125 06:41:40.896968 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.712038 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:41:48 crc kubenswrapper[4708]: E1125 06:41:48.712993 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24b001fe-8c17-4769-acd2-cbdb94abf51d" containerName="container-00" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.713010 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="24b001fe-8c17-4769-acd2-cbdb94abf51d" containerName="container-00" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.713253 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="24b001fe-8c17-4769-acd2-cbdb94abf51d" containerName="container-00" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.714640 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.725980 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.755334 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.755554 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc2rs\" (UniqueName: \"kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.755646 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.857133 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.857249 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc2rs\" (UniqueName: \"kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.857308 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.857831 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.857958 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:48 crc kubenswrapper[4708]: I1125 06:41:48.879908 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc2rs\" (UniqueName: \"kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs\") pod \"certified-operators-kxvmt\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.038578 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.513164 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.893083 4708 generic.go:334] "Generic (PLEG): container finished" podID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerID="375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e" exitCode=0 Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.893273 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerDied","Data":"375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e"} Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.893389 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerStarted","Data":"9edafd068d0a489d9c0333a47f1f8cf7709d87a56b3240e781d2e243d7066907"} Nov 25 06:41:49 crc kubenswrapper[4708]: I1125 06:41:49.895609 4708 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 06:41:51 crc kubenswrapper[4708]: I1125 06:41:51.922878 4708 generic.go:334] "Generic (PLEG): container finished" podID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerID="00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f" exitCode=0 Nov 25 06:41:51 crc kubenswrapper[4708]: I1125 06:41:51.922992 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerDied","Data":"00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f"} Nov 25 06:41:52 crc kubenswrapper[4708]: I1125 06:41:52.933432 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerStarted","Data":"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0"} Nov 25 06:41:52 crc kubenswrapper[4708]: I1125 06:41:52.951132 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kxvmt" podStartSLOduration=2.431115107 podStartE2EDuration="4.951113036s" podCreationTimestamp="2025-11-25 06:41:48 +0000 UTC" firstStartedPulling="2025-11-25 06:41:49.895365109 +0000 UTC m=+3651.304198495" lastFinishedPulling="2025-11-25 06:41:52.415363038 +0000 UTC m=+3653.824196424" observedRunningTime="2025-11-25 06:41:52.946315824 +0000 UTC m=+3654.355149210" watchObservedRunningTime="2025-11-25 06:41:52.951113036 +0000 UTC m=+3654.359946422" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.240650 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84bbd797c4-gvnft_236c83bd-a2bb-4d2b-b43b-212abfb9eb1f/barbican-api/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.313497 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84bbd797c4-gvnft_236c83bd-a2bb-4d2b-b43b-212abfb9eb1f/barbican-api-log/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.407590 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6db8477fbb-lpxf5_06962f63-df39-4943-8021-1e0f4d337962/barbican-keystone-listener/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.493649 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6db8477fbb-lpxf5_06962f63-df39-4943-8021-1e0f4d337962/barbican-keystone-listener-log/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.584440 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8dcdb5d95-6tkvf_92442175-4e83-4a5e-92ee-b75bfeca3f30/barbican-worker-log/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.609218 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-8dcdb5d95-6tkvf_92442175-4e83-4a5e-92ee-b75bfeca3f30/barbican-worker/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.777926 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-k52wj_eecde679-d675-471a-9e80-a38eca4feba7/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.851077 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/ceilometer-central-agent/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.893871 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:41:53 crc kubenswrapper[4708]: E1125 06:41:53.894135 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.946646 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/ceilometer-notification-agent/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.969492 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/proxy-httpd/0.log" Nov 25 06:41:53 crc kubenswrapper[4708]: I1125 06:41:53.986330 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_8108f47c-e809-47f5-991d-d42b19b6e8aa/sg-core/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.182467 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_607ff629-5168-4560-a9bc-c4598573405b/cinder-api-log/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.370633 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_607ff629-5168-4560-a9bc-c4598573405b/cinder-api/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.371155 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fc1392ab-0788-442a-8664-1a674811e537/cinder-scheduler/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.386261 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fc1392ab-0788-442a-8664-1a674811e537/probe/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.538711 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jtrzl_87296b4d-6e9f-4206-9d0f-78dd40e51e17/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.558120 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-wfz4n_9084cfbd-0f77-4a71-b265-fdbe8ac68bef/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.687863 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/init/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.820861 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/init/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.854966 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5d64f6bfbc-4f9hv_0f3ea896-7a11-4d35-94ae-0aa01889f4c8/dnsmasq-dns/0.log" Nov 25 06:41:54 crc kubenswrapper[4708]: I1125 06:41:54.892549 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-ws4vm_555ae52e-3696-4f22-b959-d5b9407242bf/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.040843 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0af2b359-a58e-4bc7-b497-40368df40ee1/glance-httpd/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.071821 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_0af2b359-a58e-4bc7-b497-40368df40ee1/glance-log/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.165218 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6891a73d-d8e8-4a12-9136-b56912aeeb54/glance-httpd/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.209080 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_6891a73d-d8e8-4a12-9136-b56912aeeb54/glance-log/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.254956 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-244cm_a0c3d4bb-310d-4de3-84d8-67f8654706b7/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.423873 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-4lx94_ed0b5403-c990-48b9-b3cc-8fe8455360a4/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.454748 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29400841-fnx8l_92abfbe3-d658-4e40-aa2b-38167fe164fd/keystone-cron/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.725804 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f6e8f862-6f67-426c-8e31-cb1f6baaa44e/kube-state-metrics/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.945642 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-bslnv_b24673ce-c915-4c77-aa4a-83ad370e794e/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:55 crc kubenswrapper[4708]: I1125 06:41:55.953247 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-db6d644f4-2hdvr_231fbeca-a7d9-4b28-b6ba-d1ed5ab61892/keystone-api/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.258303 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-574d769d5-pkqlp_3abea71e-bc4c-4e36-a462-bd420d9b0090/neutron-httpd/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.260725 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-574d769d5-pkqlp_3abea71e-bc4c-4e36-a462-bd420d9b0090/neutron-api/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.332876 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-65fmz_c88370ee-fc37-4d14-851f-87c2be9e9bf1/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.343861 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c2a8af8f-ba8a-43c0-84dd-d3d7f1fceaf8/memcached/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.835993 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_f6d2e9ec-0ff0-4db9-af17-51d9e824252d/nova-cell0-conductor-conductor/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.850926 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12f918b9-1449-414e-abc9-723b7e31570b/nova-api-log/0.log" Nov 25 06:41:56 crc kubenswrapper[4708]: I1125 06:41:56.983074 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_6b0f9369-64c1-47a1-83ca-052238803461/nova-cell1-conductor-conductor/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.159144 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_e9ad8e31-f124-4cc8-ab77-ff50df6589f1/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.239728 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_12f918b9-1449-414e-abc9-723b7e31570b/nova-api-api/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.240787 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-wxw56_7497674d-bdb1-427d-b4ae-cda53421f90d/nova-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.363603 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd/nova-metadata-log/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.587711 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/mysql-bootstrap/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.652538 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_815de079-46a6-4ff4-b4f3-f1e3294a21c9/nova-scheduler-scheduler/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.752291 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/mysql-bootstrap/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.771370 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_b8379446-127e-4f18-b730-085e4b80e60b/galera/0.log" Nov 25 06:41:57 crc kubenswrapper[4708]: I1125 06:41:57.896901 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/mysql-bootstrap/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.093066 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/galera/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.117474 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_9c745e5e-026a-4659-8bfd-9c8939a234f4/openstackclient/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.121590 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_33d2c4ee-3cf2-434f-bbaf-b2087a17b0bc/mysql-bootstrap/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.307578 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_c6568a70-fb7c-4b63-9f9f-dcd3af9b47cd/nova-metadata-metadata/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.359706 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bdsjf_77b982b3-1e42-4b3a-b3b9-e19f104405ed/openstack-network-exporter/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.376283 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server-init/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.510831 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server-init/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.527252 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovsdb-server/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.582910 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-xnwzt_003334fe-bdcf-4d36-8ba6-705eb6ae4fe3/ovs-vswitchd/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.615966 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-qc989_4fe9cd9f-2e87-43fa-b3c1-09560d973e2f/ovn-controller/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.727367 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-kh7hm_4b251231-3104-491a-9dc1-bac1869d716d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.797966 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_bd40d3a2-afea-42fe-93b4-72902b3046a6/ovn-northd/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.828582 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_bd40d3a2-afea-42fe-93b4-72902b3046a6/openstack-network-exporter/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.887678 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_131d2c08-5b83-48df-914b-1f5c28b0ba47/openstack-network-exporter/0.log" Nov 25 06:41:58 crc kubenswrapper[4708]: I1125 06:41:58.968991 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_131d2c08-5b83-48df-914b-1f5c28b0ba47/ovsdbserver-nb/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.027700 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4c903b56-d8a2-4522-97d1-4b2333be4929/openstack-network-exporter/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.038939 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.038984 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.051688 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4c903b56-d8a2-4522-97d1-4b2333be4929/ovsdbserver-sb/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.085164 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.238858 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8ff6575b-8w7c4_17af25e9-9b9e-4175-b35c-18fe0b9e0c36/placement-api/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.267892 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/setup-container/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.301142 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-7b8ff6575b-8w7c4_17af25e9-9b9e-4175-b35c-18fe0b9e0c36/placement-log/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.591011 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/setup-container/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.598194 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_66867ebe-f2fd-4936-80f8-dd70668c35ab/rabbitmq/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.652863 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/setup-container/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.813410 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/setup-container/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.835110 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-pfnc6_1ffad4ac-de8a-4f8c-8409-20b8553a4259/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.859012 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_74727331-8900-458f-8a20-0e8ad82d4f43/rabbitmq/0.log" Nov 25 06:41:59 crc kubenswrapper[4708]: I1125 06:41:59.979418 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-d6wtp_19981de6-1c24-415c-9d45-c2d31d8b54fd/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.044634 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.050547 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-72fph_50e3a68c-2f8a-48c7-a330-aa646ee174dd/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.093672 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.111280 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-6k7zh_1ef0663b-adb0-4192-a17a-1643899f466f/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.213183 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-t7t6t_a2cb620c-a539-4928-9a66-6c08cf197798/ssh-known-hosts-edpm-deployment/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.309131 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b4dc74dc-ktwwc_34548024-cb82-4fa9-ace7-9e41b3ab2c1d/proxy-server/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.367092 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b4dc74dc-ktwwc_34548024-cb82-4fa9-ace7-9e41b3ab2c1d/proxy-httpd/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.411011 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-p5f7m_14772fed-d981-4b42-ac9b-24dd6ce02b06/swift-ring-rebalance/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.498170 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-auditor/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.527438 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-reaper/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.596501 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-replicator/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.611059 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/account-server/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.701139 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-server/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.702426 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-replicator/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.715698 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-auditor/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.766812 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/container-updater/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.799684 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-auditor/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.867797 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-expirer/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.913225 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-replicator/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.917136 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-server/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.953588 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/object-updater/0.log" Nov 25 06:42:00 crc kubenswrapper[4708]: I1125 06:42:00.973595 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/rsync/0.log" Nov 25 06:42:01 crc kubenswrapper[4708]: I1125 06:42:01.040656 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee6df83-f664-41b7-9c47-594a0920da51/swift-recon-cron/0.log" Nov 25 06:42:01 crc kubenswrapper[4708]: I1125 06:42:01.098433 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-2ng7q_180c2306-33c0-4aed-93f8-3c99603f678b/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:42:01 crc kubenswrapper[4708]: I1125 06:42:01.193093 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_695e5c80-15cf-4dc5-b286-cb10e12e9fee/tempest-tests-tempest-tests-runner/0.log" Nov 25 06:42:01 crc kubenswrapper[4708]: I1125 06:42:01.282495 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_655d4e8f-367a-475f-8363-1506473bb0d4/test-operator-logs-container/0.log" Nov 25 06:42:01 crc kubenswrapper[4708]: I1125 06:42:01.339514 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-4w55g_ec676a79-152d-4660-854e-5e447bd52ff2/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.015540 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kxvmt" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="registry-server" containerID="cri-o://84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0" gracePeriod=2 Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.437138 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.495694 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc2rs\" (UniqueName: \"kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs\") pod \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.495794 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities\") pod \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.495834 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content\") pod \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\" (UID: \"67d91a9a-ecfe-433a-ad53-21ab8175ff7f\") " Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.496493 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities" (OuterVolumeSpecName: "utilities") pod "67d91a9a-ecfe-433a-ad53-21ab8175ff7f" (UID: "67d91a9a-ecfe-433a-ad53-21ab8175ff7f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.501603 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs" (OuterVolumeSpecName: "kube-api-access-lc2rs") pod "67d91a9a-ecfe-433a-ad53-21ab8175ff7f" (UID: "67d91a9a-ecfe-433a-ad53-21ab8175ff7f"). InnerVolumeSpecName "kube-api-access-lc2rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.544499 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67d91a9a-ecfe-433a-ad53-21ab8175ff7f" (UID: "67d91a9a-ecfe-433a-ad53-21ab8175ff7f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.597565 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc2rs\" (UniqueName: \"kubernetes.io/projected/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-kube-api-access-lc2rs\") on node \"crc\" DevicePath \"\"" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.597596 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:42:02 crc kubenswrapper[4708]: I1125 06:42:02.597608 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67d91a9a-ecfe-433a-ad53-21ab8175ff7f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.025244 4708 generic.go:334] "Generic (PLEG): container finished" podID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerID="84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0" exitCode=0 Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.025283 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerDied","Data":"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0"} Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.025316 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kxvmt" event={"ID":"67d91a9a-ecfe-433a-ad53-21ab8175ff7f","Type":"ContainerDied","Data":"9edafd068d0a489d9c0333a47f1f8cf7709d87a56b3240e781d2e243d7066907"} Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.025338 4708 scope.go:117] "RemoveContainer" containerID="84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.025479 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kxvmt" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.045099 4708 scope.go:117] "RemoveContainer" containerID="00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.045208 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.059932 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kxvmt"] Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.077836 4708 scope.go:117] "RemoveContainer" containerID="375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.101953 4708 scope.go:117] "RemoveContainer" containerID="84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0" Nov 25 06:42:03 crc kubenswrapper[4708]: E1125 06:42:03.102355 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0\": container with ID starting with 84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0 not found: ID does not exist" containerID="84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.102406 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0"} err="failed to get container status \"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0\": rpc error: code = NotFound desc = could not find container \"84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0\": container with ID starting with 84280a40a49b58348a89857634cabe4e8c8f03226ff1bbe90d6203b5c369f7d0 not found: ID does not exist" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.102431 4708 scope.go:117] "RemoveContainer" containerID="00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f" Nov 25 06:42:03 crc kubenswrapper[4708]: E1125 06:42:03.102919 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f\": container with ID starting with 00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f not found: ID does not exist" containerID="00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.102969 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f"} err="failed to get container status \"00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f\": rpc error: code = NotFound desc = could not find container \"00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f\": container with ID starting with 00e48625147a520d0c8bc81e6b9c76d605e9feb2f16db0a60720250524ff353f not found: ID does not exist" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.102992 4708 scope.go:117] "RemoveContainer" containerID="375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e" Nov 25 06:42:03 crc kubenswrapper[4708]: E1125 06:42:03.103269 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e\": container with ID starting with 375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e not found: ID does not exist" containerID="375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e" Nov 25 06:42:03 crc kubenswrapper[4708]: I1125 06:42:03.103291 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e"} err="failed to get container status \"375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e\": rpc error: code = NotFound desc = could not find container \"375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e\": container with ID starting with 375d57b340d84e091e3e5de915c91bfd95e8e984567a904c84adc784c717485e not found: ID does not exist" Nov 25 06:42:04 crc kubenswrapper[4708]: I1125 06:42:04.902563 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" path="/var/lib/kubelet/pods/67d91a9a-ecfe-433a-ad53-21ab8175ff7f/volumes" Nov 25 06:42:07 crc kubenswrapper[4708]: I1125 06:42:07.894026 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:42:07 crc kubenswrapper[4708]: E1125 06:42:07.894559 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.027245 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-kvrrg_8cb5eab2-d137-4aeb-9ca5-59139c67a2de/kube-rbac-proxy/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.090657 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-kvrrg_8cb5eab2-d137-4aeb-9ca5-59139c67a2de/manager/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.173373 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.354336 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.370785 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.411366 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.529671 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/util/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.560895 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/extract/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.563747 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_bbe0292a041351b2e91c74017e768208b36f144dd799fdf82c414fd15ffsh6g_322a79ef-4a89-4999-a433-93fc72601580/pull/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.697939 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-b7hd2_13a9d474-f528-4913-ada4-313db91a37d5/kube-rbac-proxy/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.733184 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-j2vk8_fd2a735a-3219-45d8-a689-1a8722923d8d/kube-rbac-proxy/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.734923 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-b7hd2_13a9d474-f528-4913-ada4-313db91a37d5/manager/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.874971 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-j2vk8_fd2a735a-3219-45d8-a689-1a8722923d8d/manager/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.893042 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:42:19 crc kubenswrapper[4708]: E1125 06:42:19.893454 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.919220 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-df8zr_eef4d5f6-a70a-4b08-8836-3dde6ab0da36/kube-rbac-proxy/0.log" Nov 25 06:42:19 crc kubenswrapper[4708]: I1125 06:42:19.999140 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-68b95954c9-df8zr_eef4d5f6-a70a-4b08-8836-3dde6ab0da36/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.111553 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-9knwb_e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.119331 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-9knwb_e0d5a94b-c7fe-4a06-aa42-5746d4e79ca2/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.215630 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j98rt_b752b0d3-bffb-4322-9e11-d2a9451c4f39/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.293830 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-j98rt_b752b0d3-bffb-4322-9e11-d2a9451c4f39/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.333051 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-j5g96_49cb03d2-7430-4f92-b82b-7e4b7ad9faee/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.519954 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-d5cc86f4b-j5g96_49cb03d2-7430-4f92-b82b-7e4b7ad9faee/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.522598 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-8xsh6_1fd35d81-f228-4c45-895a-95feb523ef1f/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.529758 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-8xsh6_1fd35d81-f228-4c45-895a-95feb523ef1f/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.651535 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-9wnpj_1e34dcae-a471-4c9c-83f2-43d551516520/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.754156 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-9wnpj_1e34dcae-a471-4c9c-83f2-43d551516520/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.839238 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-6zc24_b2240efc-4688-4d79-a19d-ea3801742494/kube-rbac-proxy/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.846863 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-6zc24_b2240efc-4688-4d79-a19d-ea3801742494/manager/0.log" Nov 25 06:42:20 crc kubenswrapper[4708]: I1125 06:42:20.903911 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-x4xzv_856fc8ab-76a7-4ed4-ab1e-31dad3985982/kube-rbac-proxy/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.013990 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-x4xzv_856fc8ab-76a7-4ed4-ab1e-31dad3985982/manager/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.066039 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-c8c5l_17ef75be-981e-4579-ae1b-aa7e4c4b8918/kube-rbac-proxy/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.116695 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-c8c5l_17ef75be-981e-4579-ae1b-aa7e4c4b8918/manager/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.211147 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-qjcbh_dcbf8314-b18c-43ce-8a2e-400b9da62852/kube-rbac-proxy/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.293015 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-qjcbh_dcbf8314-b18c-43ce-8a2e-400b9da62852/manager/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.366267 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rmmvr_6af5f14d-1c9e-4985-ba46-3f4b193ac96d/kube-rbac-proxy/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.398726 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-rmmvr_6af5f14d-1c9e-4985-ba46-3f4b193ac96d/manager/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.482102 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gk8ft_3d11487d-d74d-428d-b18b-800238416691/kube-rbac-proxy/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.569260 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-b58f89467-gk8ft_3d11487d-d74d-428d-b18b-800238416691/manager/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.873708 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-7b567956b5-lb9hh_5b8ebbd7-267d-4d0c-8d41-b760d18cd2f1/operator/0.log" Nov 25 06:42:21 crc kubenswrapper[4708]: I1125 06:42:21.903919 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-zkbgb_7dc7a5c4-2123-4ca4-ad42-364233f79bb5/registry-server/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.038533 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-gnn9s_a7c6b74d-a672-4349-b9f7-52a726bc8ec9/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.164367 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-gnn9s_a7c6b74d-a672-4349-b9f7-52a726bc8ec9/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.171640 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-vzz64_1064823e-17fb-42bc-b68e-cfab481fbf8c/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.288032 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-vzz64_1064823e-17fb-42bc-b68e-cfab481fbf8c/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.386979 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mc72h_54a0ed25-ded8-471d-acdd-b34f75347701/operator/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.544245 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-d64tb_83077424-8cf5-48ae-b8b2-d0b94fb650e6/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.634260 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-d64tb_83077424-8cf5-48ae-b8b2-d0b94fb650e6/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.690036 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7cd5954d9-c7dbb_af33b085-a912-42a9-88ba-6b7e8d27ec8d/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.694955 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-phd66_7d60669a-b373-49bf-b65f-88b3e0d48d87/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.782661 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-phd66_7d60669a-b373-49bf-b65f-88b3e0d48d87/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.839749 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-z9kvf_ed82966d-e1bb-448f-be61-091fcb5f3f2b/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.861139 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-z9kvf_ed82966d-e1bb-448f-be61-091fcb5f3f2b/manager/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.953138 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-hhgb4_cb75b241-90c4-45ee-9473-864b6b9f8705/kube-rbac-proxy/0.log" Nov 25 06:42:22 crc kubenswrapper[4708]: I1125 06:42:22.999033 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-hhgb4_cb75b241-90c4-45ee-9473-864b6b9f8705/manager/0.log" Nov 25 06:42:34 crc kubenswrapper[4708]: I1125 06:42:34.893509 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:42:34 crc kubenswrapper[4708]: E1125 06:42:34.894335 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:42:36 crc kubenswrapper[4708]: I1125 06:42:36.886134 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-642cw_e1dcf9d6-33d7-4482-b24f-990de83f2fa1/control-plane-machine-set-operator/0.log" Nov 25 06:42:37 crc kubenswrapper[4708]: I1125 06:42:37.007248 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f98lb_e41814f6-cdb7-4460-bc75-618b5fff39df/kube-rbac-proxy/0.log" Nov 25 06:42:37 crc kubenswrapper[4708]: I1125 06:42:37.045330 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-f98lb_e41814f6-cdb7-4460-bc75-618b5fff39df/machine-api-operator/0.log" Nov 25 06:42:45 crc kubenswrapper[4708]: I1125 06:42:45.894041 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:42:45 crc kubenswrapper[4708]: E1125 06:42:45.895030 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:42:47 crc kubenswrapper[4708]: I1125 06:42:47.659249 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-pfp8b_8775bc9d-3d1d-471a-b6af-c057f25a7f12/cert-manager-controller/0.log" Nov 25 06:42:47 crc kubenswrapper[4708]: I1125 06:42:47.817476 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-t4xc2_6cd001d7-f756-4a83-8d4b-0bc9c5198b0a/cert-manager-cainjector/0.log" Nov 25 06:42:47 crc kubenswrapper[4708]: I1125 06:42:47.889994 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-7dc5l_d043f20a-ee01-40c7-86a5-ff37f52d7f07/cert-manager-webhook/0.log" Nov 25 06:42:57 crc kubenswrapper[4708]: I1125 06:42:57.894974 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:42:57 crc kubenswrapper[4708]: E1125 06:42:57.896067 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.034790 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:42:58 crc kubenswrapper[4708]: E1125 06:42:58.035515 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="extract-content" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.035573 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="extract-content" Nov 25 06:42:58 crc kubenswrapper[4708]: E1125 06:42:58.035590 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="registry-server" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.035597 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="registry-server" Nov 25 06:42:58 crc kubenswrapper[4708]: E1125 06:42:58.035624 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="extract-utilities" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.035631 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="extract-utilities" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.035873 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="67d91a9a-ecfe-433a-ad53-21ab8175ff7f" containerName="registry-server" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.037153 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.056211 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.177598 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.177940 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67km2\" (UniqueName: \"kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.178215 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.280794 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.280834 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67km2\" (UniqueName: \"kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.280974 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.281454 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.281757 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.310355 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67km2\" (UniqueName: \"kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2\") pod \"redhat-marketplace-dwv8v\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.357652 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:42:58 crc kubenswrapper[4708]: I1125 06:42:58.808251 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.346209 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-w982m_d6725355-22e6-4236-94e9-ce851f0d8e29/nmstate-console-plugin/0.log" Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.511684 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerID="324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54" exitCode=0 Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.511733 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerDied","Data":"324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54"} Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.511760 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerStarted","Data":"c6d5ee7a60e944bd9909eb7541a813a5f23a94f308f11939f421cc8fa4d8bced"} Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.513079 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-gghd9_7590e799-a5e1-4fc1-b9ba-c3d6d9084cb7/nmstate-handler/0.log" Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.568362 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kjddh_f7336ae4-7e4b-48a3-8fbd-618e0491a2ff/kube-rbac-proxy/0.log" Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.583424 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-kjddh_f7336ae4-7e4b-48a3-8fbd-618e0491a2ff/nmstate-metrics/0.log" Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.752993 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-dvth5_daa36bec-fbbf-44f3-8785-8e09474d7add/nmstate-operator/0.log" Nov 25 06:42:59 crc kubenswrapper[4708]: I1125 06:42:59.754737 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-2xlzf_03a47fb5-3577-45b2-83b2-cedcd799eea0/nmstate-webhook/0.log" Nov 25 06:43:01 crc kubenswrapper[4708]: I1125 06:43:01.531336 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerID="54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01" exitCode=0 Nov 25 06:43:01 crc kubenswrapper[4708]: I1125 06:43:01.531431 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerDied","Data":"54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01"} Nov 25 06:43:02 crc kubenswrapper[4708]: I1125 06:43:02.553111 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerStarted","Data":"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7"} Nov 25 06:43:02 crc kubenswrapper[4708]: I1125 06:43:02.573962 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-dwv8v" podStartSLOduration=2.10875372 podStartE2EDuration="4.57394419s" podCreationTimestamp="2025-11-25 06:42:58 +0000 UTC" firstStartedPulling="2025-11-25 06:42:59.513919611 +0000 UTC m=+3720.922752997" lastFinishedPulling="2025-11-25 06:43:01.97911008 +0000 UTC m=+3723.387943467" observedRunningTime="2025-11-25 06:43:02.566929829 +0000 UTC m=+3723.975763214" watchObservedRunningTime="2025-11-25 06:43:02.57394419 +0000 UTC m=+3723.982777576" Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.357909 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.358660 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.398499 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.662575 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.714813 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:43:08 crc kubenswrapper[4708]: I1125 06:43:08.899986 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:43:08 crc kubenswrapper[4708]: E1125 06:43:08.900273 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:43:10 crc kubenswrapper[4708]: I1125 06:43:10.637159 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-dwv8v" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="registry-server" containerID="cri-o://ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7" gracePeriod=2 Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.062281 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.243867 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67km2\" (UniqueName: \"kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2\") pod \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.244168 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content\") pod \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.244238 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities\") pod \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\" (UID: \"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3\") " Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.245110 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities" (OuterVolumeSpecName: "utilities") pod "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" (UID: "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.250660 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2" (OuterVolumeSpecName: "kube-api-access-67km2") pod "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" (UID: "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3"). InnerVolumeSpecName "kube-api-access-67km2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.261818 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" (UID: "b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.347920 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67km2\" (UniqueName: \"kubernetes.io/projected/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-kube-api-access-67km2\") on node \"crc\" DevicePath \"\"" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.347960 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.347972 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.650991 4708 generic.go:334] "Generic (PLEG): container finished" podID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerID="ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7" exitCode=0 Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.651040 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerDied","Data":"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7"} Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.651070 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-dwv8v" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.651094 4708 scope.go:117] "RemoveContainer" containerID="ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.651078 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-dwv8v" event={"ID":"b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3","Type":"ContainerDied","Data":"c6d5ee7a60e944bd9909eb7541a813a5f23a94f308f11939f421cc8fa4d8bced"} Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.696743 4708 scope.go:117] "RemoveContainer" containerID="54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.704479 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.721916 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-dwv8v"] Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.771576 4708 scope.go:117] "RemoveContainer" containerID="324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.803793 4708 scope.go:117] "RemoveContainer" containerID="ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7" Nov 25 06:43:11 crc kubenswrapper[4708]: E1125 06:43:11.804459 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7\": container with ID starting with ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7 not found: ID does not exist" containerID="ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.804536 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7"} err="failed to get container status \"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7\": rpc error: code = NotFound desc = could not find container \"ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7\": container with ID starting with ea715c624ede7f7a99dc8c28a507bb20a0b1a5736fb06cb5f611a5b25d5e8fb7 not found: ID does not exist" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.804567 4708 scope.go:117] "RemoveContainer" containerID="54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01" Nov 25 06:43:11 crc kubenswrapper[4708]: E1125 06:43:11.804935 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01\": container with ID starting with 54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01 not found: ID does not exist" containerID="54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.804974 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01"} err="failed to get container status \"54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01\": rpc error: code = NotFound desc = could not find container \"54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01\": container with ID starting with 54b647e1aa2143a35513081f7b612654870fc56190b7c39fcc160853d48a1a01 not found: ID does not exist" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.805004 4708 scope.go:117] "RemoveContainer" containerID="324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54" Nov 25 06:43:11 crc kubenswrapper[4708]: E1125 06:43:11.805242 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54\": container with ID starting with 324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54 not found: ID does not exist" containerID="324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54" Nov 25 06:43:11 crc kubenswrapper[4708]: I1125 06:43:11.805271 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54"} err="failed to get container status \"324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54\": rpc error: code = NotFound desc = could not find container \"324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54\": container with ID starting with 324c0e547b2f07096100121c12e5e426810da666c33d60e4b7e11c3f3873dc54 not found: ID does not exist" Nov 25 06:43:12 crc kubenswrapper[4708]: I1125 06:43:12.902676 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" path="/var/lib/kubelet/pods/b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3/volumes" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.142259 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bkbdq_56eb3e20-e964-4e88-a2f7-8f07c3e8d40f/kube-rbac-proxy/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.275202 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bkbdq_56eb3e20-e964-4e88-a2f7-8f07c3e8d40f/controller/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.376117 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.572338 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.574240 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.577110 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.645538 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.799717 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.828400 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.835430 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:43:13 crc kubenswrapper[4708]: I1125 06:43:13.862132 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.010724 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-metrics/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.023587 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-reloader/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.036219 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/controller/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.056776 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/cp-frr-files/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.231683 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/frr-metrics/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.247677 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/kube-rbac-proxy/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.291592 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/kube-rbac-proxy-frr/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.479899 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/reloader/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.484423 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-5fbpc_f3e32da4-a772-4f5b-b91a-fbd4f5405a94/frr-k8s-webhook-server/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.667238 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5885d8f89f-gjghz_93f12b13-309e-4a76-9f10-c1ca9a0545eb/manager/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.917237 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-66489dff8c-hbstx_9ae15378-5b01-442c-a1b3-52d25e474875/webhook-server/0.log" Nov 25 06:43:14 crc kubenswrapper[4708]: I1125 06:43:14.947294 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dg7zk_fb80d906-e252-4e84-95a3-8c63ef010175/kube-rbac-proxy/0.log" Nov 25 06:43:15 crc kubenswrapper[4708]: I1125 06:43:15.298658 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-f8lcr_415c292a-1706-4c7b-9f9e-30f469708872/frr/0.log" Nov 25 06:43:15 crc kubenswrapper[4708]: I1125 06:43:15.414924 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-dg7zk_fb80d906-e252-4e84-95a3-8c63ef010175/speaker/0.log" Nov 25 06:43:22 crc kubenswrapper[4708]: I1125 06:43:22.893158 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:43:22 crc kubenswrapper[4708]: E1125 06:43:22.893946 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.295238 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.479603 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.499763 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.505435 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.628213 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/pull/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.635135 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/util/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.642217 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ebxv8j_0d46ed7f-a1de-40aa-8b62-f59e5770c6ad/extract/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.776387 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.931178 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.949406 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:43:26 crc kubenswrapper[4708]: I1125 06:43:26.969946 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.074350 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-utilities/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.082162 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/extract-content/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.267496 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.470017 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.486202 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.537832 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.578499 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-6ffgp_e85ce5fc-a263-4980-8186-4741cb651bbf/registry-server/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.630621 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-utilities/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.674092 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/extract-content/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.811411 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.979326 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:43:27 crc kubenswrapper[4708]: I1125 06:43:27.983721 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.020265 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.133727 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-9w72p_3aa64bad-9407-4512-ac3c-ab4c0d27a896/registry-server/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.214880 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/extract/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.217195 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/pull/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.221494 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c68vfpn_615985f2-6094-4f12-8ab8-c19cb9d18971/util/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.377766 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-qlzlc_7fa8126c-8308-4ce3-8529-70b28a606428/marketplace-operator/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.418248 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.550847 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.561626 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.578577 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.728011 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-utilities/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.734909 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/extract-content/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.895033 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-b8fwz_4295a1d6-dc34-4671-be69-a6c0a6b5f698/registry-server/0.log" Nov 25 06:43:28 crc kubenswrapper[4708]: I1125 06:43:28.958241 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.083270 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.094114 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.118058 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.267043 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-content/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.273810 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/extract-utilities/0.log" Nov 25 06:43:29 crc kubenswrapper[4708]: I1125 06:43:29.691200 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-w9wdd_8bf634fc-fea4-4334-9015-38cb2e07095a/registry-server/0.log" Nov 25 06:43:37 crc kubenswrapper[4708]: I1125 06:43:37.893889 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:43:37 crc kubenswrapper[4708]: E1125 06:43:37.896222 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:43:48 crc kubenswrapper[4708]: I1125 06:43:48.901360 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:43:48 crc kubenswrapper[4708]: E1125 06:43:48.902315 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:43:51 crc kubenswrapper[4708]: E1125 06:43:51.653174 4708 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 192.168.26.34:60102->192.168.26.34:46789: read tcp 192.168.26.34:60102->192.168.26.34:46789: read: connection reset by peer Nov 25 06:43:59 crc kubenswrapper[4708]: I1125 06:43:59.893219 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:43:59 crc kubenswrapper[4708]: E1125 06:43:59.894306 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:44:10 crc kubenswrapper[4708]: I1125 06:44:10.893006 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:44:10 crc kubenswrapper[4708]: E1125 06:44:10.893934 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:44:25 crc kubenswrapper[4708]: I1125 06:44:25.893488 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:44:25 crc kubenswrapper[4708]: E1125 06:44:25.894543 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:44:38 crc kubenswrapper[4708]: I1125 06:44:38.899645 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:44:38 crc kubenswrapper[4708]: E1125 06:44:38.900619 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:44:51 crc kubenswrapper[4708]: I1125 06:44:51.574304 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1f20800-2884-4fb3-ade5-bc1860682bea" containerID="af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b" exitCode=0 Nov 25 06:44:51 crc kubenswrapper[4708]: I1125 06:44:51.574398 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" event={"ID":"e1f20800-2884-4fb3-ade5-bc1860682bea","Type":"ContainerDied","Data":"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b"} Nov 25 06:44:51 crc kubenswrapper[4708]: I1125 06:44:51.577024 4708 scope.go:117] "RemoveContainer" containerID="af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b" Nov 25 06:44:51 crc kubenswrapper[4708]: E1125 06:44:51.612893 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1f20800_2884_4fb3_ade5_bc1860682bea.slice/crio-af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.513452 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4cx2b_must-gather-tsjj4_e1f20800-2884-4fb3-ade5-bc1860682bea/gather/0.log" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.893140 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:44:52 crc kubenswrapper[4708]: E1125 06:44:52.893421 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.984533 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:44:52 crc kubenswrapper[4708]: E1125 06:44:52.986559 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="extract-utilities" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.986582 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="extract-utilities" Nov 25 06:44:52 crc kubenswrapper[4708]: E1125 06:44:52.986600 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="extract-content" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.986607 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="extract-content" Nov 25 06:44:52 crc kubenswrapper[4708]: E1125 06:44:52.986645 4708 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="registry-server" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.986651 4708 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="registry-server" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.986831 4708 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d2d42b-ca20-4bd0-8b7e-e23362c85cf3" containerName="registry-server" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.988212 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:52 crc kubenswrapper[4708]: I1125 06:44:52.999229 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.003793 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klbvn\" (UniqueName: \"kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.003856 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.004168 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.105651 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klbvn\" (UniqueName: \"kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.105716 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.105776 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.106207 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.106245 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.128485 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klbvn\" (UniqueName: \"kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn\") pod \"community-operators-cvzvn\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.309209 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:44:53 crc kubenswrapper[4708]: I1125 06:44:53.814013 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:44:54 crc kubenswrapper[4708]: I1125 06:44:54.602433 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2019eab-d83d-46ff-85ac-b06a7cd7e245" containerID="c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92" exitCode=0 Nov 25 06:44:54 crc kubenswrapper[4708]: I1125 06:44:54.602481 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerDied","Data":"c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92"} Nov 25 06:44:54 crc kubenswrapper[4708]: I1125 06:44:54.602549 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerStarted","Data":"0baf8d7847b21c63a31c9e897456f1f3b916f7bf1026db407245efc5daaacf68"} Nov 25 06:44:55 crc kubenswrapper[4708]: I1125 06:44:55.614823 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerStarted","Data":"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2"} Nov 25 06:44:56 crc kubenswrapper[4708]: I1125 06:44:56.628109 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2019eab-d83d-46ff-85ac-b06a7cd7e245" containerID="fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2" exitCode=0 Nov 25 06:44:56 crc kubenswrapper[4708]: I1125 06:44:56.628176 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerDied","Data":"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2"} Nov 25 06:44:57 crc kubenswrapper[4708]: I1125 06:44:57.646543 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerStarted","Data":"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103"} Nov 25 06:44:57 crc kubenswrapper[4708]: I1125 06:44:57.669823 4708 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cvzvn" podStartSLOduration=3.170393928 podStartE2EDuration="5.669802196s" podCreationTimestamp="2025-11-25 06:44:52 +0000 UTC" firstStartedPulling="2025-11-25 06:44:54.604724271 +0000 UTC m=+3836.013557658" lastFinishedPulling="2025-11-25 06:44:57.10413254 +0000 UTC m=+3838.512965926" observedRunningTime="2025-11-25 06:44:57.664168717 +0000 UTC m=+3839.073002103" watchObservedRunningTime="2025-11-25 06:44:57.669802196 +0000 UTC m=+3839.078635582" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.175063 4708 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p"] Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.177399 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.181302 4708 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.181335 4708 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.181470 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p"] Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.261433 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.261639 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.261666 4708 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rbv6\" (UniqueName: \"kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.363170 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.363251 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rbv6\" (UniqueName: \"kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.363304 4708 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.364569 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.379946 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rbv6\" (UniqueName: \"kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.382981 4708 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume\") pod \"collect-profiles-29400885-tnd2p\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.498291 4708 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:00 crc kubenswrapper[4708]: I1125 06:45:00.872954 4708 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p"] Nov 25 06:45:00 crc kubenswrapper[4708]: W1125 06:45:00.875188 4708 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf274399a_84d4_400c_9689_5ed04286aca5.slice/crio-1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266 WatchSource:0}: Error finding container 1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266: Status 404 returned error can't find the container with id 1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266 Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.696242 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4cx2b/must-gather-tsjj4"] Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.696861 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" podUID="e1f20800-2884-4fb3-ade5-bc1860682bea" containerName="copy" containerID="cri-o://3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec" gracePeriod=2 Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.708737 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4cx2b/must-gather-tsjj4"] Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.714377 4708 generic.go:334] "Generic (PLEG): container finished" podID="f274399a-84d4-400c-9689-5ed04286aca5" containerID="2f3b923b5b736972d38a90427b978a3eeeb96354a96f54bf715554f29e44efe1" exitCode=0 Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.714414 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" event={"ID":"f274399a-84d4-400c-9689-5ed04286aca5","Type":"ContainerDied","Data":"2f3b923b5b736972d38a90427b978a3eeeb96354a96f54bf715554f29e44efe1"} Nov 25 06:45:01 crc kubenswrapper[4708]: I1125 06:45:01.714436 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" event={"ID":"f274399a-84d4-400c-9689-5ed04286aca5","Type":"ContainerStarted","Data":"1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266"} Nov 25 06:45:01 crc kubenswrapper[4708]: E1125 06:45:01.871902 4708 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1f20800_2884_4fb3_ade5_bc1860682bea.slice/crio-3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1f20800_2884_4fb3_ade5_bc1860682bea.slice/crio-conmon-3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec.scope\": RecentStats: unable to find data in memory cache]" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.066572 4708 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4cx2b_must-gather-tsjj4_e1f20800-2884-4fb3-ade5-bc1860682bea/copy/0.log" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.067488 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.201826 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output\") pod \"e1f20800-2884-4fb3-ade5-bc1860682bea\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.202105 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmvvc\" (UniqueName: \"kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc\") pod \"e1f20800-2884-4fb3-ade5-bc1860682bea\" (UID: \"e1f20800-2884-4fb3-ade5-bc1860682bea\") " Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.207722 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc" (OuterVolumeSpecName: "kube-api-access-pmvvc") pod "e1f20800-2884-4fb3-ade5-bc1860682bea" (UID: "e1f20800-2884-4fb3-ade5-bc1860682bea"). InnerVolumeSpecName "kube-api-access-pmvvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.305266 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmvvc\" (UniqueName: \"kubernetes.io/projected/e1f20800-2884-4fb3-ade5-bc1860682bea-kube-api-access-pmvvc\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.317255 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "e1f20800-2884-4fb3-ade5-bc1860682bea" (UID: "e1f20800-2884-4fb3-ade5-bc1860682bea"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.407279 4708 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/e1f20800-2884-4fb3-ade5-bc1860682bea-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.724703 4708 generic.go:334] "Generic (PLEG): container finished" podID="e1f20800-2884-4fb3-ade5-bc1860682bea" containerID="3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec" exitCode=143 Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.724762 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4cx2b/must-gather-tsjj4" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.724787 4708 scope.go:117] "RemoveContainer" containerID="3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.748324 4708 scope.go:117] "RemoveContainer" containerID="af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.792889 4708 scope.go:117] "RemoveContainer" containerID="3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec" Nov 25 06:45:02 crc kubenswrapper[4708]: E1125 06:45:02.793306 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec\": container with ID starting with 3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec not found: ID does not exist" containerID="3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.793348 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec"} err="failed to get container status \"3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec\": rpc error: code = NotFound desc = could not find container \"3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec\": container with ID starting with 3ae53435413dad8b62fd08eae59f60ef5b502be17dfe05c500dca5a408e012ec not found: ID does not exist" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.793372 4708 scope.go:117] "RemoveContainer" containerID="af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b" Nov 25 06:45:02 crc kubenswrapper[4708]: E1125 06:45:02.793735 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b\": container with ID starting with af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b not found: ID does not exist" containerID="af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.793769 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b"} err="failed to get container status \"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b\": rpc error: code = NotFound desc = could not find container \"af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b\": container with ID starting with af2f95df828ab96e259e787fde050c6c7ef90f5dc06c191c5b19aa304b97424b not found: ID does not exist" Nov 25 06:45:02 crc kubenswrapper[4708]: I1125 06:45:02.903322 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f20800-2884-4fb3-ade5-bc1860682bea" path="/var/lib/kubelet/pods/e1f20800-2884-4fb3-ade5-bc1860682bea/volumes" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.004343 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.018701 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rbv6\" (UniqueName: \"kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6\") pod \"f274399a-84d4-400c-9689-5ed04286aca5\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.025705 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6" (OuterVolumeSpecName: "kube-api-access-6rbv6") pod "f274399a-84d4-400c-9689-5ed04286aca5" (UID: "f274399a-84d4-400c-9689-5ed04286aca5"). InnerVolumeSpecName "kube-api-access-6rbv6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.121394 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume\") pod \"f274399a-84d4-400c-9689-5ed04286aca5\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.121455 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume\") pod \"f274399a-84d4-400c-9689-5ed04286aca5\" (UID: \"f274399a-84d4-400c-9689-5ed04286aca5\") " Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.122179 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rbv6\" (UniqueName: \"kubernetes.io/projected/f274399a-84d4-400c-9689-5ed04286aca5-kube-api-access-6rbv6\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.128663 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f274399a-84d4-400c-9689-5ed04286aca5" (UID: "f274399a-84d4-400c-9689-5ed04286aca5"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.139201 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume" (OuterVolumeSpecName: "config-volume") pod "f274399a-84d4-400c-9689-5ed04286aca5" (UID: "f274399a-84d4-400c-9689-5ed04286aca5"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.223820 4708 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f274399a-84d4-400c-9689-5ed04286aca5-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.224102 4708 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f274399a-84d4-400c-9689-5ed04286aca5-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.309804 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.309855 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.348608 4708 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.735980 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.735972 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29400885-tnd2p" event={"ID":"f274399a-84d4-400c-9689-5ed04286aca5","Type":"ContainerDied","Data":"1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266"} Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.736039 4708 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bc6f2d68a246d92d45022f6d7db38bc416fb2df179eec3aa87d079e7cbad266" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.773088 4708 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:03 crc kubenswrapper[4708]: I1125 06:45:03.814045 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:45:04 crc kubenswrapper[4708]: I1125 06:45:04.101211 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc"] Nov 25 06:45:04 crc kubenswrapper[4708]: I1125 06:45:04.107958 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29400840-w85zc"] Nov 25 06:45:04 crc kubenswrapper[4708]: I1125 06:45:04.902751 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b6feecd-35f4-494c-beb1-a570abc0b382" path="/var/lib/kubelet/pods/3b6feecd-35f4-494c-beb1-a570abc0b382/volumes" Nov 25 06:45:05 crc kubenswrapper[4708]: I1125 06:45:05.752319 4708 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cvzvn" podUID="a2019eab-d83d-46ff-85ac-b06a7cd7e245" containerName="registry-server" containerID="cri-o://80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103" gracePeriod=2 Nov 25 06:45:05 crc kubenswrapper[4708]: I1125 06:45:05.893057 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:45:05 crc kubenswrapper[4708]: E1125 06:45:05.893329 4708 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-lh62q_openshift-machine-config-operator(f24da5a0-0c94-4513-9c44-84818f240046)\"" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" podUID="f24da5a0-0c94-4513-9c44-84818f240046" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.698578 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.772696 4708 generic.go:334] "Generic (PLEG): container finished" podID="a2019eab-d83d-46ff-85ac-b06a7cd7e245" containerID="80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103" exitCode=0 Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.772740 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerDied","Data":"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103"} Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.772767 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvzvn" event={"ID":"a2019eab-d83d-46ff-85ac-b06a7cd7e245","Type":"ContainerDied","Data":"0baf8d7847b21c63a31c9e897456f1f3b916f7bf1026db407245efc5daaacf68"} Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.772768 4708 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvzvn" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.772797 4708 scope.go:117] "RemoveContainer" containerID="80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.789406 4708 scope.go:117] "RemoveContainer" containerID="fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.813285 4708 scope.go:117] "RemoveContainer" containerID="c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.843594 4708 scope.go:117] "RemoveContainer" containerID="80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103" Nov 25 06:45:06 crc kubenswrapper[4708]: E1125 06:45:06.844165 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103\": container with ID starting with 80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103 not found: ID does not exist" containerID="80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.844198 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103"} err="failed to get container status \"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103\": rpc error: code = NotFound desc = could not find container \"80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103\": container with ID starting with 80017ce18207031319d829d1f91764593577fb675193b628774a155d88b12103 not found: ID does not exist" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.844216 4708 scope.go:117] "RemoveContainer" containerID="fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2" Nov 25 06:45:06 crc kubenswrapper[4708]: E1125 06:45:06.844405 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2\": container with ID starting with fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2 not found: ID does not exist" containerID="fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.844427 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2"} err="failed to get container status \"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2\": rpc error: code = NotFound desc = could not find container \"fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2\": container with ID starting with fc6ff77d5627598f662c6e1fdd8d49240e9f82c0964ab66ca2d9848f70c277f2 not found: ID does not exist" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.844441 4708 scope.go:117] "RemoveContainer" containerID="c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92" Nov 25 06:45:06 crc kubenswrapper[4708]: E1125 06:45:06.844648 4708 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92\": container with ID starting with c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92 not found: ID does not exist" containerID="c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.844671 4708 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92"} err="failed to get container status \"c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92\": rpc error: code = NotFound desc = could not find container \"c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92\": container with ID starting with c2a5f77c8e83209c327871345039144359296314a2363eb7ac2532cbffc88a92 not found: ID does not exist" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.899700 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities\") pod \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.900023 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content\") pod \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.900082 4708 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klbvn\" (UniqueName: \"kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn\") pod \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\" (UID: \"a2019eab-d83d-46ff-85ac-b06a7cd7e245\") " Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.900845 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities" (OuterVolumeSpecName: "utilities") pod "a2019eab-d83d-46ff-85ac-b06a7cd7e245" (UID: "a2019eab-d83d-46ff-85ac-b06a7cd7e245"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.903778 4708 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.905603 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn" (OuterVolumeSpecName: "kube-api-access-klbvn") pod "a2019eab-d83d-46ff-85ac-b06a7cd7e245" (UID: "a2019eab-d83d-46ff-85ac-b06a7cd7e245"). InnerVolumeSpecName "kube-api-access-klbvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 06:45:06 crc kubenswrapper[4708]: I1125 06:45:06.942722 4708 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2019eab-d83d-46ff-85ac-b06a7cd7e245" (UID: "a2019eab-d83d-46ff-85ac-b06a7cd7e245"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 06:45:07 crc kubenswrapper[4708]: I1125 06:45:07.004446 4708 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klbvn\" (UniqueName: \"kubernetes.io/projected/a2019eab-d83d-46ff-85ac-b06a7cd7e245-kube-api-access-klbvn\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:07 crc kubenswrapper[4708]: I1125 06:45:07.004690 4708 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2019eab-d83d-46ff-85ac-b06a7cd7e245-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 06:45:07 crc kubenswrapper[4708]: I1125 06:45:07.104764 4708 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:45:07 crc kubenswrapper[4708]: I1125 06:45:07.111818 4708 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cvzvn"] Nov 25 06:45:08 crc kubenswrapper[4708]: I1125 06:45:08.903822 4708 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2019eab-d83d-46ff-85ac-b06a7cd7e245" path="/var/lib/kubelet/pods/a2019eab-d83d-46ff-85ac-b06a7cd7e245/volumes" Nov 25 06:45:18 crc kubenswrapper[4708]: I1125 06:45:18.898491 4708 scope.go:117] "RemoveContainer" containerID="12a3cfd6050b3c8be1d7c4ffad34d1d2d8f046b2163d7df81666d358a7783dd4" Nov 25 06:45:19 crc kubenswrapper[4708]: I1125 06:45:19.876943 4708 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-lh62q" event={"ID":"f24da5a0-0c94-4513-9c44-84818f240046","Type":"ContainerStarted","Data":"67af1a5c9996dc18aec8a5cc409487f90661e9b843376ea479e27a214b6cfaf9"} Nov 25 06:46:02 crc kubenswrapper[4708]: I1125 06:46:02.714570 4708 scope.go:117] "RemoveContainer" containerID="5ea71a3d3a99497eb4cb5799a305177661bbdf49770a539a5d396e8d08d71778" Nov 25 06:47:02 crc kubenswrapper[4708]: I1125 06:47:02.799754 4708 scope.go:117] "RemoveContainer" containerID="7098dc40bd9845ff9772139c42740cb807d895964f4b39e1c3a30113a8ea19a8" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111250364024443 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111250364017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111240337016502 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111240340015444 5ustar corecore